textLSP
textLSP copied to clipboard
Language server for text spell and grammar check with various tools.
textLSP
Language server for text spell and grammar check with various AI tools.
This tool is in early development.
Features
LSP features
- Diagnostics:
- spelling or grammatical errors
- Code actions:
- Fix suggestions
- Analyze paragraph with a selected passive analyzer (if the analyzer does not check on save or change)
Showcase
- Only on the first character of the first line: analyze the whole document if it was not fully checked yet
Showcase
- Custom actions defined by a given analyzer
E.g. OpenAI text generation
- Context based word suggestion
Showcase
Analyzers
Local tools
The following tools run on the local system:
- LanguageTool: Mainly for development purposes, see ltex-ls for a more mature implementation.
-
Gramformer: Neural network based system.
-
Gramformer needs to be installed manually:
pip install git+https://github.com/PrithivirajDamodaran/Gramformer.git
-
- hf_checker: Huggingface
text2text-generation
pipline based analyser. See the flan-t5-large-grammar-synthesis model for an example.Models
- pszemraj/grammar-synthesis-small
- pszemraj/grammar-synthesis-large
- pszemraj/flan-t5-large-grammar-synthesis
- pszemraj/flan-t5-xl-grammar-synthesis
- pszemraj/bart-base-grammar-synthesis
- hf_instruction_checker: Huggingface
text2text-generation
pipline based analyser using instruction tuned models. See the Grammarly's CoEdIT model for an example. Supports error checking and text generation, such as paraphrasing, through the%HF%
magic command (see the OpenAI analyser below).Models
- grammarly/coedit-large
- grammarly/coedit-xl
- grammarly/coedit-xl-composite
- grammarly/coedit-xxl
- jbochi/coedit-base
- jbochi/coedit-small
- jbochi/candle-coedit-quantized
-
hf_completion: Huggingface
fill-mask
pipline based text completion.
Tools using remote services
DISCLAIMER: THE RELATED APIS REQUIRE REGISTRATION AND ARE NOT FREE TO USE! USE THESE ANALYZERS ON YOUR OWN RESPONSIBILITY! THE AUTHORS OF TEXTLSP DO NOT ASSUME ANY RESPONSIBILITY FOR THE COSTS INCURRED!
The following tools use remote text APIs. Due to potential costs turning off automatic analysis if suggested.
-
OpenAI: Supports text correction as well as text generation through a magic command in the text file.
Generation showcase
- GrammarBot: The GrammarBot API provides spelling and grammar checking.
Supported File Types
- latex
- org
- markdown
- any other file types as plain text
Setup
Install
pip install textLSP
For the latest version:
pip install git+https://github.com/hangyav/textLSP
Running
Simply run:
textlsp
Since some analyzers are computation intensive, consider running it on a server using the TCP interface:
textlsp --address 0.0.0.0 --port 1234
or simply over ssh (with ssh key) if the client doesn't support it:
ssh <server> textlsp
Configuration
Using textLSP within an editor depends on the editor of choice. For a few examples how to set up language servers in general in some of the popular editors see here or take a look at the related documentation of your editor.
By default, all analyzers are disabled in textLSP, they have to be turned on in the settings. Example configuration in lua for nvim (other editors should be set up accordingly):
textLSP = {
analysers = {
languagetool = {
enabled = true,
check_text = {
on_open = true,
on_save = true,
on_change = false,
}
},
gramformer = {
-- gramformer dependency needs to be installed manually
enabled = false,
gpu = false,
check_text = {
on_open = false,
on_save = true,
on_change = false,
}
},
hf_checker = {
enabled = false,
gpu = false,
quantize=32,
model='pszemraj/flan-t5-large-grammar-synthesis',
min_length=40,
check_text = {
on_open = false,
on_save = true,
on_change = false,
}
},
hf_instruction_checker = {
enabled = true,
gpu = false,
quantize=32,
model='grammarly/coedit-large',
min_length=40,
check_text = {
on_open = false,
on_save = true,
on_change = false,
}
},
hf_completion = {
enabled = true,
gpu = false,
quantize=32,
model='bert-base-multilingual-cased',
topk=5,
},
openai = {
enabled = false,
api_key = '<MY_API_KEY>',
check_text = {
on_open = false,
on_save = false,
on_change = false,
},
model = 'gpt-3.5-turbo',
max_token = 16,
},
grammarbot = {
enabled = false,
api_key = '<MY_API_KEY>',
-- longer texts are split, this parameter sets the maximum number of splits per analysis
input_max_requests = 1,
check_text = {
on_open = false,
on_save = false,
on_change = false,
}
},
},
documents = {
-- the language of the documents, could be set to `auto` of `auto:<fallback>`
-- to detect automatically, default: auto:en
language = "auto:en",
-- do not autodetect documents with fewer characters
min_length_language_detect = 20,
org = {
org_todo_keywords = {
'TODO',
'IN_PROGRESS',
'DONE'
},
},
txt = {
parse = true,
},
},
}