Skip to content

Milestone July 2018

Sajjad edited this page Jun 23, 2018 · 23 revisions

Extraction group

Work package Name Use Case - QALD 7 multilingual train dataset
ID 0
Summary In this use case, the user of SASK will be able to explore the QALD 7 dataset (https://github.com/ag-sc/QALD/blob/master/7/data/qald-7-train-multilingual.json). The extraction components should be able to create an own KB (QALD-KB) from the abstracts attached to each resource (http://dbpedia.org/resource/XYZ) in the QALD 7 dataset. The QA and keyword search approaches should be able to work over the QALD-KB and be able to answer >10% of the queries in QALD 7. The Intelligent Data Assisstant should be able to use the property-object information from the QALD-KB to cluster the entities.
Start Date 2018-06-05
End Date 2018-07-16
Tasks Task 0.0 SASK should be able to load and extract RDF via ensemble learning from the QALD 7 entity abstracts. Therefore a corpus of abstracts should be created from the list of all resources in QALD 7 (via one SPARQL query).
Task 0.1 SASK should be able to answer questions such as "What is the capital of Cameroon?
Task 0.2 SASK should be able to visualize simple connections such as which cities share the same country.
Task 0.3 SASK should be able to answer keyword queries "movies, directed, Francis Ford Coppola".
Task 0.4 SASK should be able to create a table from certain entities and the properties they have.
Task 0.5 SASK should be able to cluster all places via the intelligent data assistant.
Deliverable Deliverable 0 - A working demo of the aforementioned tasks which feels natural.
Work package Name  Complex workflows
ID 1
Summary The application should be able to handle more complex workflows
Start Date 2018-04-23
End Date 2018-07-16
Tasks Task 1.1 (Lead: Andrè, Participants: Kevin) The backend (executer) should be able to handle more complex workflows
Task 1.2 (Lead: Kevin) The frontend (webclient) should be able to create more complex workflows
Deliverable Deliverable 1.1 (Task 1.1, Lead: Andrè, Due Date: 2018-07-16, Type: Software in the Master)
Deliverable 1.1 (Task 1.2, Lead: Kevin, Due Date: 2018-07-16, Type: Software in the Master)
Work package Name Exception & Message handling
ID 2
Summary The application should log everything on a central location
Start Date 2018-04-23
End Date 2018-07-16
Tasks Task 2.1 (Lead: Sepide, Participants: Suganya) Setup Logstash Docker
Task 2.2 (Lead: Sepide, Participants: Suganya) Setup Kibana Docker
Task 2.3 (Lead: Sepide, Participants: Suganya) Setup Logging microservice
Deliverable Deliverable 2.1 (Task 2.1, Lead: Sepide, Due Date: 2018-07-16, Type: Documentation) A wiki page with a guide to start logstash
Deliverable 2.2 (Task 2.2, Lead: Sepide, Due Date: 2018-07-16, Type: Documentation) A wiki page with a guide to start Kibana
Deliverable 2.2 (Task 2.2, Lead: Sepide, Due Date: 2018-07-16, Type: Software in Master) A new microservice to collect all logs and use lostash
Work package Name Integrate chat in UI
ID 3
Summary The chat should be intregrated in the UI
Start Date 2018-04-23
End Date 2018-07-16
Tasks Task 3.1 (Lead: Nilanjan, Participants: Kevin) create a (paper) prototype with the integrated chat window
Task 3.2 (Lead: Nilanjan, Participants: Kevin) create a functional prototype with the integrated chat window
Task 3.3 (Lead: Nilanjan, Participants: Kevin) integrate the chat in the ui
Deliverable Deliverable 3.1 (Task 3.1, Lead: Nilanjan, Due Date: 2018-05-14 Type: Document) A digital or paper prototype
Deliverable 3.2 (Task 3.2, Lead: Nilanjan, Due Date: 2018-06-04, Type: Software) A functional prototype of the new UI
Deliverable 3.3 (Task 3.3, Lead: Nilanjan, Due Date: 2018-07-16, Type: Software in Master) The chat should be integrated in the UI
Work package Name Ensemble learning
ID 4
Summary To extract data from training data To implement ensemble learning to predict which extractor fit
Start Date 2018-04-23
End Date 2018-07-16
Tasks Task( 4.1 Lead: Harsh, Participant : Kevin) To extract data from training data
Task( 4.2 Lead: Harsh, Participant : Kevin) To implement ensemble learning version for testing
Task( 4.3 Lead: Harsh, Participant : Kevin) To integrate into system
Deliverable Deliverable 4.1 (Task 4.1, Lead: Harsh, ,Due Date: 2018-07-16, Type: Document)Extracted Data
Deliverable 4.2 (Task 4.2, Lead: Harsh, Due Date: 2018-07-16, Type: Software) Ensemble Learning Software
Deliverable 4.3(Task 4.3, Lead: Harsh,Due Date: 2018-07-16, Type: Software in Master)Ensemble Learning implemented in our software
Work package Name Visualize the database content
ID 5
Summary The user should be able to display database content as graph and table in the UI
Start Date 2018-04-23
End Date 2018-07-16
Tasks Task 5.1 (Lead: Nilanjan, Participants: Kevin) Display database content in the UI
Deliverable Deliverable 5.1 (Task 5.1, Lead: Nilanjan, Due Date: 2018-07-16 Type: Software in Master)
Work package Name Auto Index Refinements
ID 6
Summary Auto Index should be highly customizable from both configuration and Rest interfaces, Close pending issues
Start Date 2018-04-23
End Date 2018-07-16
Tasks Task1: (Lead: Prashanth, Participants: Idrees) Customize Labels , Add Unit tests , Single Surface form and other bug fixes Task2: (Lead: Idrees) Quality Improvement(Codacy/Better Code) and Unit test coverage
Deliverable Due Date: 2018-07-16 Type: Software in Master which is very highly flexible and customisable)
Work package Name Intent Classification
ID 7
Summary Weka tool must be able to predict which component to delegate the query to
Start Date 2018-04-23
End Date 2018-07-16
Tasks (Lead: Divya) 1. Add Weka Code which predicts classification from a training set 2. Incorporate user feedback to improve future predictions
Deliverable Due Date: 2018-07-16 Type: Software in Master
Work package Name Deployment of Chatbot SASK Component
ID 8
Summary Come up with a generic framework for deployment of Chatbot microservices
Start Date 2018-04-23
End Date 2018-07-16
Tasks (Lead: Guru) 1. Dockerize the miroservices 2. Write a ansible playbook deploying the chatbot related microservices
Deliverable Due Date: 2018-07-16 Type: DevOps software in Master
Work package Name CSV Files Merger
ID 9
Summary Python/Java based logic to merge multiple csv files together to support visualization/clustering algorithms
Start Date 2018-04-23
End Date 2018-06-01
Tasks (Lead: Faisal)
1. Going through the Panama Papers CSV files to look for common issues while merging.
2. Writing logic to merge and store multiple files.
Deliverable Due Date: 2018-06-01 Type: Software in Master
Work package Name Missing/Null Values Identifier + Handler
ID 10
Summary Python based logic to identify missing value in a given csv file and perform mean/median/substitution operations on the missing feature values
Start Date 2018-04-23
End Date 2018-06-01
Tasks (Lead: Aditya)
1. Logic to identify missing value for given csv and feature.
2. To perform mean/median/substitution operations on the given file and feature.
Deliverable Due Date: 2018-06-01 Type: Software in Master
Work package Name Implementing visualization
ID 11
Summary Implementing visualization
Start Date 2018-04-23
End Date 2018-06-01
Tasks (Lead: Nikit )
1. Writing code snippets in Python to generate the required data in Json format for Visualization (Bar Graphs/Spring Loaded Graphs).
2. Passing the data to GUI and displaying the visualization using D3js.
Deliverable Due Date: 2018-06-01 Type: Software in Master
Work package Name Cluster Algorithm Code snippets
ID 12
Summary Cluster Algorithm Code snippets
Start Date 2018-04-23
End Date 2018-06-01
Tasks (Lead: Anu )
1. Going through list of Clustering algorithms in scikit data dump.
2. Creation of a Python based code snippet to execute the given clustering algorithm alongwith the provided parameters.
Deliverable Due Date: 2018-06-01 Type: Software in Master
Work package Name Jupyter Kernel Gateway
ID 13
Summary Configuring a Jupyter Kernel Gateway
Start Date 2018-06-03
End Date 2018-06-08
Tasks (Lead: Nikit)
1. Setting up methods on Java for api calls to Jupyter Kernel Gateway. Designed in a way where user can create/end sessions and execute python code snippets
Deliverable Due Date: 2018-06-08 Type: Software in Master
Work package Name Inter-Task communication handling
ID 14
Summary Inter-Task communication handling
Start Date 2018-04-23
End Date 2018-06-22
Tasks (Lead: Nikit )
1. Making a list of requirements/output of each task and ensuring smooth transition between the workflow steps, discussing the solutions and resolution of conflicts between tasks
Deliverable Due Date: 2018-06-22 Type: Software in Master
Work package Name Rivescript Storyboard creation
ID 15
Summary Rivescript Storyboard creation
Start Date 2018-06-03
End Date 2018-06-22
Tasks (Lead: Juzer, Participants: Anu, Aditya Bhat, Faisal, Nikit )
1. Communicating responsible person of each task and collecting the requirements for the creation of Rivescript storyboard
Deliverable Due Date: 2018-06-22 Type: Software in Master
Work package Name Integration of Components
ID 16
Summary Integration of the components for Data Science Workflow to create a working prototype.
Start Date 2018-06-22
End Date 2018-07-06
Tasks (Lead: Nikit, Participants: Anu, Aditya Bhat, Faisal, Juzer )
1. Integrating the Storyboard Narrative and Code Snippets into a single working prototype.
Deliverable Due Date: 2018-07-06 Type: Software in Master
Work package Name Handle SASK Workflow
ID 17
Summary Handle SASK Workflow
Start Date 2018-04-23
End Date 2018-06-22
Tasks (Lead: Juzer )
1. Handling the workflow from Intelligent Data Assistant to the related modules
Deliverable Due Date: 2018-06-22 Type: Software in Master
Work package Name Study SurniaQA and create tasks
ID 18
Summary Study existing code and groom next development tasks
Start Date 2018-05-28
End Date 2018-07-12
Tasks 1.Set up development environment and clone SurniaQA repo. 2. Testing of SASK 3. Setup SurniaQA with autoindex 4. Create microservice to talk to SurniaQA (Sajjad & Nesara)
Deliverable Due Date: 2018-07-12 Type: Tasks on trello/git for next development cycle
Clone this wiki locally