Download presentation
Presentation is loading. Please wait.
Published byDonald Hunter Modified over 9 years ago
1
Automating programming via concept mining, probabilistic reasoning over semantic knowledge base of SE domain by Max Talanov
2
A lot of trivial tasks that could be automated Add/remove field Patronymic on Customer page Add dropdown list on the form... A lot of not so trivial solutions that should be reused but are not reused How-tos Libraries... Approximately 60% of developer's time in outsourcing is spent to solve this kind of problems </ Problem
3
Template based code generators: IDEs Visual studio. IDEA. ... Template based generators Maven Archetypes CASE Tools: Rational Rose ArgoUML </ Current Solutions
4
Once generated solution is hardly maintainable and require to allocate significant amount of money for farther support. Developer generated the solution based on DB structure, then added some functionality in it, then customer wants to change the DB structure. Developer has to regenerate and merge his further changes This is done for only one reason – generator does not understand what it's doing. </ They cannot solve problem
5
2005 MIT media lab published the article “FEASIBILITY STUDIES FOR PROGRAMMING IN NATURAL LANGUAGE”. Metafor is the program that creates the sceleton of the Python classes based on shallow English descripton. </ MIT Metafor
6
Input: shallow English description Output: scaffolding Python classes. Metafor utilised natural language processor Montylingua, common sence KB ConceptNet, programmatic interpreter. </ MIT Metafor
7
System should: Operate with changes to be applied not the static structure of the target application (How-to). Use domain knowledge model to map the inbound requirements into acceptance criteria to be used to create the solution. Use both trained data and generation to create the solution. Use several abstraction layers of the target application. </ Key Ideas
8
System has to operate with knowledge. System has to understand what it is doing Architecture of the target application. Methods to change the architecture. Domain specific information. Requirements for changes. System has to understand the human operator Communicate in natural language. </ Requirements
9
Knowledge base. Linguistic component. Perceiving component. Solution generator. Communication component. </ Key Components
10
Linguistic Req Perceiving Solution generator Communicator Updated app Request KB </ Collaboration Diagram
11
Is the main storage of the data to be used by the system. KB is RDF storage with OWL data. KB is used to store the semantic information: Target application architecture Domain specific knowledge (How-tos) Common sense information Predicates generated based on requirements text Acceptance criteria for generated solution ... </ Knowledge Base
12
Human expert specifies the requirements, linguistuc component generates set of predicates for further processing. Inbound: CR, bug report or FRS according to SE standard SPICE. Outbound: set of predicates. Stanford Parser creates the set of predicates that are treated as inbound knowledge. Linguistic Requirements </ Linguistic Component
13
Perceiving module maps inbound predicates to the model (domain model) in knowledge base, using trained data and stochastic search generation. In case of failure invokes Communicator to generate clarification request. Inbound: set of predicates and domain knowledge model. Outbound: predicates mapped to domain knowledge model. LinguisticPerceiving Solution generator Predicates Updated model </ Perceiving Component
14
Selects or generates the solution for the specified acceptance criteria in updated model. Provides the updated application with the confirmation request to human expert. Inbound: acceptance criteria in KB Outbound: solution in actual code. Updated app Perceiving Solution generator Updated model </ Solution Generator Components
15
Reasoner. Reasoner interface. Genetic generator. Solution checker. Trainer. Associator. Generalizer. Analogy detector. Target language translator. </ Solution Generator Components
16
</ Solution Generator Activities Analogy Detector: retrieves solution for similar acceptance criteria Solution checker: returns solution assessment Genetic generator: generates new solution Communicator: sends confirmation request to human expert. Communicator: analyses reply of human expert Trainer: run Solutuon found? Solution ok? yes no
17
For the purpose of making a logical inference for axepted alternatives within environment of possible contradictions and several probable variants, we decided to use probabilistic reasoner NARS. NARS main features: Deduction Induction Analogy ... </ Reasoner
18
Inbound: Acceptance criteria. Outbound: Solution in form of How-tos sequences. Generator is capable of creation of the sequences of How-tos according to the acceptance criteria. Inference is produced by NARS. This could be interpreted as human imagination mechanism. </ Genetic Generator
19
Is devoted to logically infer the percentage of how good is generated solution according to acceptance criteria. Solution checker mainly relies on NARS probabilistic mechanisms, but collects all proper information from KB to be processed by reasoner. </ Solution Checker
20
Machine learning component, is used to detect associations and infer generic associations of inbound acceptance criteria and approved solutions. Analogy detector is used to retrieve previously learned associations that could be used for specified acceptance criteria. </ Trainer
21
Translates knowledge representation of architecture in actual files in target language, based on previously described syntax in KB. </ Target Language Translator
22
Component is dedicated to generate requests to human expert and analyse replyes of the expert. Perceiving Solution generator Communicator Request </ Communicator
23
</ Feedback Loops Linguistic Req Perceiving Solution generator Communicator Updated app Request KB
24
</ Current Implementation Acceptance criteria Solution generator Communicator Updated app Request KB
25
Add evolution mechanism. Add inbound information analysis. Add architectural analysis. Add self optimization and self improvement. Extend perceiving algorithm to use encyclopaedic resources to extend domain knowledge. </ Future Plans
26
Metafor : http://web.media.mit.edu/~lieber/Publications/Feasibility-Nat-Lang- Prog.pdf Maven : http://maven.apache.org/guides/introduction/introduction-to-the- lifecycle.html Stanford Parser: http://nlp.stanford.edu/software/lex-parser.shtml Open NARS: http://code.google.com/p/open-nars/ Menta: http://code.google.com/p/menta/ </ References
Similar presentations
© 2024 SlidePlayer.com. Inc.
All rights reserved.