File: doc.go

package info (click to toggle)
golang-github-shenwei356-bio 0.13.3-1
  • links: PTS, VCS
  • area: main
  • in suites: forky, sid, trixie
  • size: 852 kB
  • sloc: perl: 114; sh: 58; makefile: 21
file content (90 lines) | stat: -rw-r--r-- 2,166 bytes parent folder | download | duplicates (3)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
/*Package fastx seamlessly parses FASTA and FASTQ format file
This package seamlessly parses both FASTA and FASTQ formats.

## Examples

### Common operation

    package main

    import (
    	"fmt"
    	"io"
    	"os"

    	// "github.com/shenwei356/bio/seq"
    	"github.com/shenwei356/bio/seqio/fastx"
    	"github.com/shenwei356/xopen"
    )

    func main() {
    	// use buffered out stream for output
    	outfh, err := xopen.Wopen("-") // "-" for STDOUT
    	checkError(err)
    	defer outfh.Close()

    	// disable sequence validation could reduce time when reading large sequences
    	// seq.ValidateSeq = false
        
    	reader, err := fastx.NewDefaultReader("-")
    	checkError(err)
    	for {
    		record, err := reader.Read()
    		if err != nil {
    			if err == io.EOF {
    				break
    			}
    			checkError(err)
    			break
    		}

    		// fmt is slow for output, because it's not buffered
    		// fmt.Printf("%s", record.Format(0))

    		record.FormatToWriter(outfh, 0)
    	}
    }

    func checkError(err error) {
    	if err != nil {
    		fmt.Fprintln(os.Stderr, err)
    		os.Exit(1)
    	}
    }


***Note that***, similar with `bytes.Buffer.Bytes()` method,
the current record will change after your another call of this method.
You may use `record.Clone()` to make a copy.

### Asynchronously parsing

`ChunkChan` asynchronously reads FASTA/Q records, and returns a channel of
Record Chunk, from which you can easily access the records.
`bufferSize` is the number of buffered chunks, and `chunkSize` is the size
of records in a chunk.

    reader, err := fastx.NewDefaultReader(file)
    checkError(err)

    for chunk := range reader.ChunkChan(bufferSize, chunkSize) {
        checkError(chunk.Err)

        for _, record := range chunk.Data {
            fmt.Print(record)
        }
    }

***Note that***, these's no need to clone the record by `record.Clone()` here.

### Custom alphabet and identifier regular expression

    import (
        "github.com/shenwei356/bio/seq"
        "github.com/shenwei356/bio/seqio/fastx"
    )

    reader, err := fastx.NewReader(seq.DNA, file, "^([^\s]+)\s?")

*/
package fastx