The aim of the bachelor thesis is to implement a framework for inertial motion capture during music performance. The musician plays an instrument (e.g. guitar) and the movement of his arms, hands and fingers is being captured. In our solution, the inertial-based motion capture glove and HTC Vive tracker, are used to track the arms and hands. Afterwards, the animation produced from the inertial motion capture data is applied to a 3D character model, which can be rendered in virtual reality. In addition to the inertial devices, the whole performance is captured by a 3D camera and an animated point cloud is generated. Another goal of the thesis is to synchronize the inertial data with the optical data obtained from the camera. In further research, the synchronization can be used when capturing the movement using only the optical methods, where the inertial motion capture will be used as a ground truth to evaluate the skeleton.