Planned maintenance
A system upgrade is planned for 24/9-2024, at 12:00-14:00. During this time DiVA will be unavailable.
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Energy optimal scheduling by reinforcement learning
University West, Department of Engineering Science, Division of Production Systems.
2020 (English)Independent thesis Advanced level (degree of Master (Two Years)), 20 credits / 30 HE creditsStudent thesis
Abstract [en]

While scheduling is extensively studied by operations research and heuristic methods, both paradigms face challenges to include uncertainties into the problem and to be used on a realtime basis. Reinforcement learning is a well-documented approach for solving various planning and control problems. In this thesis, application of reinforcement learning for production scheduling is studied. Inverse scheduling is defined as the problem of finding the number of jobs and duration times to meet an input capacity demand. A reinforcement learning framework is proposed for solving inverse scheduling while optimizing energy consumption. Discrete event simulation is used to model the environment while involving uncertainties. Due to the inherent characteristics of the problem in hand, deep neural networks are applied for approximating the policy. The trained agent can be used both for scheduling (production planning) and rescheduling (production control). A production cell is studied as a testbed, and using the trained agent, a six percent reduction in the energy level is observed in simulation following a proposed energy signature. Strengths and weaknesses of the suggested framework are presented, and essential features identifying its success are discussed. To generalize the approach to large scale problems, further modifications are needed both at the side of modelling and implementation. The thesis is conducted as part of a joint project between University West, Chalmers University of Technology, and Volvo Group Trucks Operations, named SmoothIT.

Place, publisher, year, edition, pages
2020. , p. 86
Keywords [en]
Job-shop scheduling, Reinforcement learning, Energy optimization, Discrete event simulation.
National Category
Robotics
Identifiers
URN: urn:nbn:se:hv:diva-15419Local ID: EXC915OAI: oai:DiVA.org:hv-15419DiVA, id: diva2:1453283
Subject / course
Robotics
Educational program
Produktionsteknik
Supervisors
Examiners
Available from: 2020-07-22 Created: 2020-07-09 Last updated: 2020-07-22Bibliographically approved

Open Access in DiVA

No full text in DiVA

By organisation
Division of Production Systems
Robotics

Search outside of DiVA

GoogleGoogle Scholar

urn-nbn

Altmetric score

urn-nbn
Total: 535 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf