Dashboard
Introduction
Get started with Front, Responsive Website Template for building responsive, mobile-first sites, with Bootstrap and a template starter page.
Front is a Static HTML Template
For the time being, we do NOT offer any tutorials or any other materials on how to integrate Front with any CMS, Web Application Frameworks (including Angular, Vue, React, etc.) or any other similar technology. However, since Front is a static HTML/CSS and JS template, then it should be compatible with any backend technology and frameworks.File Structure
front
-
src
-
assets
- css - Compiled CSS files
- img - Image files
- json - JSON files
- js - Core Javascript and library wrapper files
- scss - SASS (SCSS) source files
- svg - SVG files
- svg-src - SVG source files where variables can be passed (files will be generated into svg folder)
- vendor - Third pary libraries (plugins)
- partials - HTML partials, learn more details on Front's Gulp Documentation page
- snippets - Ready-to-use combined components
- documentation - Documentation pages
- demo-* - Demo options with inner pages, such as Courses, App Marketplace etc.
- favicon.ico
- index.html
- ...
-
- dist - Generated distribution files
- gulpfiles - Gulp Toolkit files
-
build- Generated performance ready fully production files (by default the folder is not included) -
node_modules- NPM dependencies (by default the folder is not included) - config.js
- gulpfile.js
- package-lock.json
- package.json
- README.md
Starter template
Starter template is a snippet code for blank HTML page. Use the below snippet as a way to quickly start any new blank page. If you are using Front's Gulp Toolkit, you may use HTML (Gulp) snippet code (read more about it on Gulp page).
JavaScript structure
Core JavaScript
The foundation of the JavaScript structure in Front is based on one main object which does not change the root when the new functionalities are added, but instead, it only extends without affecting the behavior of the core object. The name of the object is HSCore and the content of this object looks like this:
Essentials of HS data-attributes
data-hs-*-options - Using the date-attribute, you can completely specify the settings for all plug-in parameters (except for functions) that are in the official documentation. Special cases will be described in the documentation of the corresponding wrappers/plugins. *- name of the wrapper/plugin.
Parameter names must be enclosed in double quotation marks "". "param": ...
For strings, quotation marks are required.
"stringParam": "Test string",
"hexParam": "#ff0000"
For numbers, quotation marks are optional. "intParam": 10
For boolean values, quotation marks can lead to not obvious consequences (due to implicit type conversion). It is recommended that you specify Boolean values without quotation marks. "boolParam": true
For arrays and objects - quotation marks can lead to errors, this does not apply to elements of arrays and objects, which can be simple types (see the description for simple types above).
Advantages
- Avoiding the probabilities of conflicts between Front codes and third party plugins (libraries).
- Intuitive clear architecture.
- Everything is structured, each component in its own file and in its component in the main object.
- The ability of extending functionality without affecting the behavior of the core object and not changing the existing functionality.
- Creation of wrapper components simply solves complicated initializations structures for the users.
- Very easy access to any starters components and core settings from anywhere in the template.
Tracking Progress in Natural Language Processing
Table of contents
English
- Automatic speech recognition
- CCG
- Common sense
- Constituency parsing
- Coreference resolution
- Data-to-Text Generation
- Dependency parsing
- Dialogue
- Domain adaptation
- Entity linking
- Grammatical error correction
- Information extraction
- Intent Detection and Slot Filling
- Language modeling
- Lexical normalization
- Machine translation
- Missing elements
- Multi-task learning
- Multi-modal
- Named entity recognition
- Natural language inference
- Part-of-speech tagging
- Question answering
- Relation prediction
- Relationship extraction
- Semantic textual similarity
- Semantic parsing
- Semantic role labeling
- Sentiment analysis
- Shallow syntax
- Simplification
- Stance detection
- Summarization
- Taxonomy learning
- Temporal processing
- Text classification
- Word sense disambiguation
Vietnamese
- Dependency parsing
- Machine translation
- Named entity recognition
- Part-of-speech tagging
- Word segmentation
Hindi
Chinese
For more tasks, datasets and results in Chinese, check out the Chinese NLP website.
French
Russian
Spanish
Portuguese
Korean
Nepali
Bengali
Persian
This document aims to track the progress in Natural Language Processing (NLP) and give an overview of the state-of-the-art (SOTA) across the most common NLP tasks and their corresponding datasets.
It aims to cover both traditional and core NLP tasks such as dependency parsing and part-of-speech tagging as well as more recent ones such as reading comprehension and natural language inference. The main objective is to provide the reader with a quick overview of benchmark datasets and the state-of-the-art for their task of interest, which serves as a stepping stone for further research. To this end, if there is a place where results for a task are already published and regularly maintained, such as a public leaderboard, the reader will be pointed there.
If you want to find this document again in the future, just go to nlpprogress.com
or nlpsota.com in your browser.
Contributing
Guidelines
Results Results reported in published papers are preferred; an exception may be made for influential preprints.
Datasets Datasets should have been used for evaluation in at least one published paper besides the one that introduced the dataset.
Code We recommend to add a link to an implementation
if available. You can add a Code column (see below) to the table if it does not exist.
In the Code column, indicate an official implementation with Official.
If an unofficial implementation is available, use Link (see below).
If no implementation is available, you can leave the cell empty.
Adding a new result
If you would like to add a new result, you can just click on the small edit button in the top-right corner of the file for the respective task (see below).

This allows you to edit the file in Markdown. Simply add a row to the corresponding table in the same format. Make sure that the table stays sorted (with the best result on top). After you’ve made your change, make sure that the table still looks ok by clicking on the “Preview changes” tab at the top of the page. If everything looks good, go to the bottom of the page, where you see the below form.

Add a name for your proposed change, an optional description, indicate that you would like to “Create a new branch for this commit and start a pull request”, and click on “Propose file change”.
Adding a new dataset or task
For adding a new dataset or task, you can also follow the steps above. Alternatively, you can fork the repository. In both cases, follow the steps below:
- If your task is completely new, create a new file and link to it in the table of contents above.
- If not, add your task or dataset to the respective section of the corresponding file (in alphabetical order).
- Briefly describe the dataset/task and include relevant references.
- Describe the evaluation setting and evaluation metric.
- Show how an annotated example of the dataset/task looks like.
- Add a download link if available.
- Copy the below table and fill in at least two results (including the state-of-the-art)
for your dataset/task (change Score to the metric of your dataset). If your dataset/task
has multiple metrics, add them to the right of
Score. - Submit your change as a pull request.
| Model | Score | Paper / Source | Code |
|---|---|---|---|
Wish list
These are tasks and datasets that are still missing:
- Bilingual dictionary induction
- Discourse parsing
- Keyphrase extraction
- Knowledge base population (KBP)
- More dialogue tasks
- Semi-supervised learning
- Frame-semantic parsing (FrameNet full-sentence analysis)
Exporting into a structured format
You can extract all the data into a structured, machine-readable JSON format with parsed tasks, descriptions and SOTA tables.
The instructions are in structured/README.md.
Instructions for building the site locally
Instructions for building the website locally using Jekyll can be found here.
See and talk to your users and leads immediately by importing your data into the Front Dashboard platform.
-
Import users from:
-
Capsule
Users -
Mailchimp
Users -
Webdev
Users - Or you can sync data to Front Dashboard to ensure your data is always up-to-date.