Statements in which the resource exists.
SubjectPredicateObjectContext
pubmed-article:17412559rdf:typepubmed:Citationlld:pubmed
pubmed-article:17412559lifeskim:mentionsumls-concept:C0023185lld:lifeskim
pubmed-article:17412559lifeskim:mentionsumls-concept:C0025663lld:lifeskim
pubmed-article:17412559lifeskim:mentionsumls-concept:C0035007lld:lifeskim
pubmed-article:17412559lifeskim:mentionsumls-concept:C0336537lld:lifeskim
pubmed-article:17412559lifeskim:mentionsumls-concept:C1705938lld:lifeskim
pubmed-article:17412559lifeskim:mentionsumls-concept:C1527178lld:lifeskim
pubmed-article:17412559pubmed:issue6lld:pubmed
pubmed-article:17412559pubmed:dateCreated2007-8-20lld:pubmed
pubmed-article:17412559pubmed:abstractTextAnimals' rhythmic movements, such as locomotion, are considered to be controlled by neural circuits called central pattern generators (CPGs), which generate oscillatory signals. Motivated by this biological mechanism, studies have been conducted on the rhythmic movements controlled by CPG. As an autonomous learning framework for a CPG controller, we propose in this article a reinforcement learning method we call the "CPG-actor-critic" method. This method introduces a new architecture to the actor, and its training is roughly based on a stochastic policy gradient algorithm presented recently. We apply this method to an automatic acquisition problem of control for a biped robot. Computer simulations show that training of the CPG can be successfully performed by our method, thus allowing the biped robot to not only walk stably but also adapt to environmental changes.lld:pubmed
pubmed-article:17412559pubmed:languageenglld:pubmed
pubmed-article:17412559pubmed:journalhttp://linkedlifedata.com/r...lld:pubmed
pubmed-article:17412559pubmed:citationSubsetIMlld:pubmed
pubmed-article:17412559pubmed:statusMEDLINElld:pubmed
pubmed-article:17412559pubmed:monthAuglld:pubmed
pubmed-article:17412559pubmed:issn0893-6080lld:pubmed
pubmed-article:17412559pubmed:authorpubmed-author:MoriTakeshiTlld:pubmed
pubmed-article:17412559pubmed:authorpubmed-author:IshiiShinSlld:pubmed
pubmed-article:17412559pubmed:authorpubmed-author:NakamuraYutak...lld:pubmed
pubmed-article:17412559pubmed:authorpubmed-author:SatoMasa-akiM...lld:pubmed
pubmed-article:17412559pubmed:issnTypePrintlld:pubmed
pubmed-article:17412559pubmed:volume20lld:pubmed
pubmed-article:17412559pubmed:ownerNLMlld:pubmed
pubmed-article:17412559pubmed:authorsCompleteYlld:pubmed
pubmed-article:17412559pubmed:pagination723-35lld:pubmed
pubmed-article:17412559pubmed:meshHeadingpubmed-meshheading:17412559...lld:pubmed
pubmed-article:17412559pubmed:meshHeadingpubmed-meshheading:17412559...lld:pubmed
pubmed-article:17412559pubmed:meshHeadingpubmed-meshheading:17412559...lld:pubmed
pubmed-article:17412559pubmed:meshHeadingpubmed-meshheading:17412559...lld:pubmed
pubmed-article:17412559pubmed:meshHeadingpubmed-meshheading:17412559...lld:pubmed
pubmed-article:17412559pubmed:year2007lld:pubmed
pubmed-article:17412559pubmed:articleTitleReinforcement learning for a biped robot based on a CPG-actor-critic method.lld:pubmed
pubmed-article:17412559pubmed:affiliationNara Institute of Science and Technology, 8916-5 Takayama-cho, Ikoma, Nara 630-0192, Japan. nakamura@ams.eng.osaka-u.ac.jplld:pubmed
pubmed-article:17412559pubmed:publicationTypeJournal Articlelld:pubmed
pubmed-article:17412559pubmed:publicationTypeResearch Support, Non-U.S. Gov'tlld:pubmed