1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73
|
package lexers
import (
"embed"
"io/fs"
"github.com/alecthomas/chroma/v2"
)
//go:embed embedded
var embedded embed.FS
// GlobalLexerRegistry is the global LexerRegistry of Lexers.
var GlobalLexerRegistry = func() *chroma.LexerRegistry {
reg := chroma.NewLexerRegistry()
// index(reg)
paths, err := fs.Glob(embedded, "embedded/*.xml")
if err != nil {
panic(err)
}
for _, path := range paths {
reg.Register(chroma.MustNewXMLLexer(embedded, path))
}
return reg
}()
// Names of all lexers, optionally including aliases.
func Names(withAliases bool) []string {
return GlobalLexerRegistry.Names(withAliases)
}
// Get a Lexer by name, alias or file extension.
func Get(name string) chroma.Lexer {
return GlobalLexerRegistry.Get(name)
}
// MatchMimeType attempts to find a lexer for the given MIME type.
func MatchMimeType(mimeType string) chroma.Lexer {
return GlobalLexerRegistry.MatchMimeType(mimeType)
}
// Match returns the first lexer matching filename.
func Match(filename string) chroma.Lexer {
return GlobalLexerRegistry.Match(filename)
}
// Register a Lexer with the global registry.
func Register(lexer chroma.Lexer) chroma.Lexer {
return GlobalLexerRegistry.Register(lexer)
}
// Analyse text content and return the "best" lexer..
func Analyse(text string) chroma.Lexer {
return GlobalLexerRegistry.Analyse(text)
}
// PlaintextRules is used for the fallback lexer as well as the explicit
// plaintext lexer.
func PlaintextRules() chroma.Rules {
return chroma.Rules{
"root": []chroma.Rule{
{`.+`, chroma.Text, nil},
{`\n`, chroma.Text, nil},
},
}
}
// Fallback lexer if no other is found.
var Fallback chroma.Lexer = chroma.MustNewLexer(&chroma.Config{
Name: "fallback",
Filenames: []string{"*"},
Priority: -1,
}, PlaintextRules)
|