# very first character of the line, i.e. without any preceeding whitespace.
-## Section 1: dlib face detection and webcam parameters
-
-# Path to the dlib shape predictor trained dataset
-predictorPath ./shape_predictor_68_face_landmarks.dat
-
-# Value passed to the cv::VideoCapture() ctor
-cvVideoCaptureId 0
-
-# Number of milliseconds to wait after processing each video frame
-# This value controls the frame rate, but the actual frame period
-# is longer due to the time required to process each frame
-cvWaitKeyMs 5
-
-# If 1, show the webcam captured video on-screen; if 0, don't show
-showWebcamVideo 1
-
-# If 1, draw the detected facial landmarks on-screen; if 0, don't draw
-renderLandmarksOnVideo 1
-
-# If 1, laterally invert the image (create a mirror image); if 0, don't invert
-lateralInversion 1
-
-
-## Section 2: Cubism params calculation control
+## Section 1: Cubism params calculation control
#
# These values control how the facial landmarks are translated into
# parameters that control the Cubism model, and will vary from person
# to person. The following values seem to work OK for my face, but
# your milage may vary.
-# Section 2.0: Live2D automatic functionality
+# If 1, laterally invert the image (create a mirror image); if 0, don't invert
+lateralInversion 1
+
+# Section 1.0: Live2D automatic functionality
# Set 1 to enable, 0 to disable.
# If these are set, the automatic functionality in Live2D will be enabled.
# Note: If you set auto blink, eye control will be disabled.
autoBreath 0
randomMotion 0
-# Section 2.1: Face Y direction angle (head pointing up/down)
+# Section 1.1: Face Y direction angle (head pointing up/down)
# The Y angle is calculated mainly based on the angle formed
# by the corners and the tip of the nose (hereafter referred
# to as the "nose angle").
faceYAngleSmileCorrection 0.075
-# Section 2.2: Eye control
+# Section 1.2: Eye control
# This is mainly calculated based on the eye aspect ratio (eye height
# divided by eye width). Note that currently an average of the values
# of both eyes is applied - mainly due to two reasons: (1) the dlib
eyeSmileMouthOpenThreshold 0.5
-# Section 2.3: Mouth control
+# Section 1.3: Mouth control
# Two parameters are passed to Cubism to control the mouth:
# - mouth form: Controls smiles / laughs
# - mouth openness: How widely open the mouth is
mouthOpenLaughCorrection 0.2
-## Section 3: Filtering parameters
+## Section 2: Filtering parameters
# The facial landmark coordinates can be quite noisy, so I've applied
# a simple moving average filter to reduce noise. More taps would mean
# more samples to average over, hence smoother movements with less noise,