All notable changes to this project will be documented in this file. This project adheres to Semantic Versioning starting with version 0.7.0.
[Unreleased 0.12.0.aX] - master
Note
This version is not yet released and is under active development.
- support for inline entity synonyms in markdown training format
- support for regex features in markdown training format
- support for splitting and training data into multiple and mixing formats
- support for markdown files containing regex-features or synonyms only
- added ability to list projects in cloud storage services for model loading
- server evaluation endpoint at
POST /evaluate
- Regex features are now sorted internally. retrain your model if you use regex features
- The keyword intent classifier now returns
null
instead of"None"
as intent name in the json result if there's no match - in teh evaluation results, replaced
O
with the stringno_entity
for better understanding - The
CRFEntityExtractor
now only trains entity examples that have"extractor": "ner_crf"
or no extractor at all - Ignore hidden files when listing projects or models
- Changelog doc formatting
- fixed project loading for newly added projects to a running server
- non ascii character support for anything that gets json dumped (e.g. training data received over HTTP endpoint)
- evaluation of entity extraction performance in
evaluation.py
- support for spacy 2.0
- evaluation of intent classification with crossvalidation in
evaluation.py
- support for splitting training data into multiple files (markdown and JSON only)
- removed
-e .
from requirements files - if you want to install the app usepip install -e .
- fixed http duckling parsing for non
en
languages - fixed parsing of entities from markdown training data files
- support asterisk style annotation of examples in markdown format
- Preventing capitalized entities from becoming synonyms of the form lower-cased -> capitalized
- read token in server from config instead of data router
- fixed reading of models with none date name prefix in server
- docker image build
- support for new dialogflow data format (previously api.ai)
- improved support for custom components (components are stored by class name in stored metadata to allow for components that are not mentioned in the Rasa NLU registry)
- language option to convert script
- Fixed loading of default model from S3. Fixes #633
- fixed permanent training status when training fails #652
- quick fix for None "_formatter_parser" bug
- readme issues
- improved setup py welcome message
- Support for training data in Markdown format
- Cors support. You can now specify allowed cors origins within your configuration file.
- The HTTP server is now backed by Klein (Twisted) instead of Flask. The server is now asynchronous but is no more WSGI compatible
- Improved Docker automated builds
- Rasa NLU now works with projects instead of models. A project can be the basis for a restaurant search bot in German or a customer service bot in English. A model can be seen as a snapshot of a project.
- Root project directories have been slightly rearranged to clean up new docker support
- use
Interpreter.create(metadata, ...)
to create interpreter from dict andInterpreter.load(file_name, ...)
to create interpreter with metadata from a file - Renamed
name
parameter toproject
- Docs hosted on GitHub pages now: Documentation
- Adapted remote cloud storages to support projects (backwards incompatible!)
- Fixed training data persistence. Fixes #510
- Fixed UTF-8 character handling when training through HTTP interface
- Invalid handling of numbers extracted from duckling during synonym handling. Fixes #517
- Only log a warning (instead of throwing an exception) on misaligned entities during mitie NER
- removed unnecessary ClassVar import
- removed obsolete
--output
parameter oftrain.py
. use--path
instead. fixes #473
- increased test coverage to avoid regressions (ongoing)
- added regex featurization to support intent classification
and entity extraction (
intent_entity_featurizer_regex
)
- replaced existing CRF library (python-crfsuite) with sklearn-crfsuite (due to better windows support)
- updated to spacy 1.8.2
- logging format of logged request now includes model name and timestamp
- use module specific loggers instead of default python root logger
- output format of the duckling extractor changed. the
value
field now includes the complete value from duckling instead of just text (so this is an property is an object now instead of just text). includes granularity information now. - deprecated
intent_examples
andentity_examples
sections in training data. all examples should go into thecommon_examples
section - weight training samples based on class distribution during ner_crf cross validation and sklearn intent classification training
- large refactoring of the internal training data structure and pipeline architecture
- numpy is now a required dependency
- luis data tokenizer configuration value (not used anymore, luis exports char offsets now)
- properly update coveralls coverage report from travis
- persistence of duckling dimensions
- changed default response of untrained
intent_classifier_sklearn
from"intent": None
to"intent": {"name": None, "confidence": 0.0}
/status
endpoint showing all available models instead of only those whose name starts with model- properly return training process ids #391
- fixed missing argument attribute error
- updated mitie installation documentation
- fixed documentation about training data format
- properly handle response_log configuration variable being set to
null
/status
endpoint showing all available models instead of only those whose name starts with model
- Fixed range calculation for crf #355
- Fixed duckling dimension persistence. fixes #358
- Fixed pypi installation dependencies (e.g. flask). fixes #354
- Fixed CRF model training without entities. fixes #345
- Fixed Luis emulation and added test to catch regression. Fixes #353
- deepcopy of context #343
- NER training reuses context inbetween requests
- ngram character featurizer (allows better handling of out-of-vocab words)
- replaced pre-wired backends with more flexible pipeline definitions
- return top 10 intents with sklearn classifier #199
- python type annotations for nearly all public functions
- added alternative method of defining entity synonyms
- support for arbitrary spacy language model names
- duckling components to provide normalized output for structured entities
- Conditional random field entity extraction (Markov model for entity tagging, better named entity recognition with low and medium data and similarly well at big data level)
- allow naming of trained models instead of generated model names
- dynamic check of requirements for the different components & error messages on missing dependencies
- support for using multiple entity extractors and combining results downstream
unified tokenizers, classifiers and feature extractors to implement common component interface
src
directory renamed torasa_nlu
when loading data in a foreign format (api.ai, luis, wit) the data gets properly split into intent & entity examples
- Configuration:
- added
max_number_of_ngrams
- removed
backend
and addedpipeline
as a replacement - added
luis_data_tokenizer
- added
duckling_dimensions
- added
- parser output format changed
from
{"intent": "greeting", "confidence": 0.9, "entities": []}
to
{"intent": {"name": "greeting", "confidence": 0.9}, "entities": []}
- entities output format changed
from
{"start": 15, "end": 28, "value": "New York City", "entity": "GPE"}
to
{"extractor": "ner_mitie", "processors": ["ner_synonyms"], "start": 15, "end": 28, "value": "New York City", "entity": "GPE"}
where
extractor
denotes the entity extractor that originally found an entity, andprocessor
denotes components that alter entities, such as the synonym component.
camel cased MITIE classes (e.g.
MITIETokenizer
→MitieTokenizer
)model metadata changed, see migration guide
updated to spacy 1.7 and dropped training and loading capabilities for the spacy component (breaks existing spacy models!)
introduced compatibility with both Python 2 and 3
- properly parse
str
additionally tounicode
#210 - support entity only training #181
- resolved conflicts between metadata and configuration values #219
- removed tokenization when reading Luis.ai data (they changed their format) #241
- fixed failed loading of example data after renaming attributes, i.e. "KeyError: 'entities'"
- fixed regression in mitie entity extraction on special characters
- fixed spacy fine tuning and entity recognition on passed language instance
- python documentation about calling rasa NLU from python
- mitie tokenization value generation #207, thanks @cristinacaputo
- changed log file extension from
.json
to.log
, since the contained text is not proper json
This is a major version update. Please also have a look at the Migration Guide.
- Changelog ;)
- option to use multi-threading during classifier training
- entity synonym support
- proper temporary file creation during tests
- mitie_sklearn backend using mitie tokenization and sklearn classification
- option to fine-tune spacy NER models
- multithreading support of build in REST server (e.g. using gunicorn)
- multitenancy implementation to allow loading multiple models which share the same backend
- error propagation on failed vector model loading (spacy)
- escaping of special characters during mitie tokenization