Hand landmark dataset. 4 days ago · Hand landmark model bundle.

Hand landmark dataset. js model on custom datasets or pre-existing ones (e.
Hand landmark dataset We annotate the real-world images with 21 landmarks and use projected ground-truth 3D joints for synthetic images. The following steps guide you through training HandSegNet and PoseNet on the Rendered Hand Pose Dataset (RHD). By leveraging the MediaPipe framework for hand landmark detection and TensorFlow for model training, this system captures live webcam video, detects hand poses, and predicts gesture categories. py # program that saves the landmarks of the hand in a numpy array along with their images # press R to The hand gesture recognition system for opening applications in Windows is a computer vision application that uses a webcam and machine learning models to recognize hand gestures and open specific 231 open source Anticlockwise-Backward-Clockwise images. However, there are only several works that address facial landmark detection in thermal images. txt files The deaf and hard-of-hearing community uses sign language for communication and interaction with the external world. 0, 1. Domain-specific datasets with labeled hand landmarks, as well as improvements to loss functions, optimizers, and architectures have pushed the latencies of gesture classifiers down to the tens of milliseconds and Oct 2, 2023 · Here is the code that helped me save the landmark positions: # hand_landmark_dataset_maker. 6M labeled single and interacting hand frames. The 21 hand landmarks are defined as follows: Apr 24, 2024 · When loading the dataset, run the pre-packaged hand detection model from MediaPipe Hands to detect the hand landmarks from the images. We extracted 21 key points from each x, y, and z, so the result is 63 key points. For this, I have designed models that detect the hand box and 9 landmarks of the hand in jpg/png images. Dataset was taken by a camera directly with white background and closeup Here are the steps to run hand landmark detection using MediaPipe. The dataset can be used for landmark recognition and retrieval experiments. Data Extraction: Implementing a two-step process involving a palm detector and a hand landmark model. Points of correspondence are placed on each image so the data set can be readily used for building statistical models of shape. 1 Input From Dataset The 6 Hand Gesture dataset contains 3066 images and 6 classes labeled Gesture1-Gesture6. The HaGRID dataset has been expanded with 15 new gesture classes, including two-handed gestures; New class "no_gesture" with domain-specific natural hand postures was addad (2,164 samples, divided by train/val/test containing 1,464, 200, 500 images, respectively) Hand Gesture Classification is a Python project that uses computer vision and machine learning to classify hand gestures in real-time. The hand landmark model bundle detects the keypoint localization of 21 hand-knuckle coordinates within the detected hand regions. # WORLD_LANDMARKS shares the same landmark topology as LANDMARKS. There are several ways to train your own hand gesture detection system. We introduce a large image dataset HaGRID (HAnd Gesture Recognition Image Dataset) for hand gesture recognition (HGR) systems. Example output of MediaPipe Hands. Each video has duration of ~1 minute (at 25-30 fps). Sign language recognition has been an active area of research for many years, and there has been progress in both sensor-based and vision-based methods. 2D key points: 2D keypoints obtained from MediaPipe for a head and for a hand face_alignment. Understanding the Hand Landmark Model. Dynamic Model Training: Train a TensorFlow. MediaPipe by default detects the (x, y, z) coordinates of 21 landmarks on the hand. In this model, we used YOLOv7 as the architecture. mediapipeHandDetection. Train Model using SVM with Hand Landmark Feature Extraction from Mediapipe with Distracted Driving Datasets Sample - kyun4/mediapipesvm Jun 4, 2022 · Firstly, a set of hand landmark points and feature were defined based on the human hand geometry and anatomy. 1. Besides a bounding box, BlazePalm also predicts 21 3D keypoints for hand landmarks (5 fingers x 4 keypoints + 1 wrist) Recognition of Hand Gestures Observed by Depth Cameras 2015, Kapuscinski et al. For more information about the capabilities, models, and configuration options of this task, see the Overview. blob: This is the model to detect the hand landmarks using the palm detection model. MediaPipe-Hand-Detection: Optimized for Mobile Deployment Real-time hand detection optimized for mobile and edge The MediaPipe Hand Landmark Detector is a machine learning pipeline that predicts bounding boxes and pose skeletons of hands in an image. 2. These landmarks are composed of the x- and y-coordinates in the image space of 21 key points of the hand, along with an additional z-coordinate; the depth of each landmark is Jun 18, 2020 · We present a real-time on-device hand tracking pipeline that predicts hand skeleton from single RGB camera for AR/VR applications. , sign_mnist). pose, left_hand, right_hand) landmark_idx : int16; x/y/z : double; Each sample corresponds to 1,629 rows (543 Label famous (and not-so-famous) landmarks in images Hand Landmarks Detected Using Google Mediapipe Gesture Recognizer Number Gestures 1-5: Hand Landmark Dataset | Kaggle Kaggle uses cookies from Google to deliver and enhance the quality of its services and to analyze traffic. r. (2) 3. The human hands can be regarded as a general-purpose interaction tool due to their dexterous functionality in communication and manipulation. The dataset was presented in our CVPR'20 paper and Google AI blog post. Some dataset used existing images from other dataset, in which case the dataset was named after the image dataset. io Hand Keypoint Detection in Single Images using Multiview Bootstrapping (Dataset) Tomas Simon, Hanbyul Joo, Iain Matthews, Yaser Sheikh. May 5, 2023 · This is an official release of InterHand2. while extracting hand landmark i encountered some valuses as None(for eg- if wrist is not visible then the wrist landmark is none) This is the second version of the Google Landmarks dataset (GLDv2), which contains images annotated with labels representing human-made and natural landmarks. The generation code 1 is made available to facilitate future studies. For each person, we annotate 4 types of bounding boxes (person box, face box, left-hand box, and right-hand box) and 133 keypoints (17 for body, 6 for feet, 68 for face and 42 for hands). We will only use static input, so we will ignore j and z, which are dynamic gestures Creating the dataset. cvlab-stonybrook/BodyHands • • CVPR 2022 We also introduce a new challenging dataset called BodyHands containing unconstrained images with hand and their corresponding body locations annotations. 2. ipynb. Images were pre-processed for two operations: preparing the original image training set and extracting the hand landmarks. Aug 12, 2024 · The datasets for these competitions are now public on the Kaggle platform. Robotic Manipulation: Enabling precise control of robotic hands. However, the infinite shapes and orientations that hands can adopt, their variability in skin pigmentation and the self-occlusions that continuously appear in images make hand segmentation a truly complex problem, especially with We try using MediaPipe's hand landmark detection model to get the numerical coordinates of the hands in each frame (which are then fed to the LSTM) as the feature extractor. 0]. The planned functionality was to store a list of lists for all xyz coordinates for the 21 hand landmark hand landmark detection. 75 PAPERS • NO BENCHMARKS YET Mar 2, 2024 · However, all datasets are used for training the hand landmark model. Otherwise, a lightweight hand tracking algorithm determines the location of the hand(s) for subsequent landmark detections. As the hand landmark points only concentrate on the 23 landmark points, issues faced can easily be avoided for image dataset such as complex background, light conditions, etc. Any images without detected hands are ommitted from the dataset. While considerable success has been achieved in 2D human landmark detection or pose estimation, there is a notable scarcity of reported works on landmark detection in unordered 3D point clouds. Apr 14, 2020 · In the end, the training set consists of the CMU Hand DB, the Egohands dataset and my own trained images (mainly from marathon runners), called cross-hands. Each head is trained by correspondent datasets marked in the same color. google. Numerous methodologies were introduced to achieve accurate and efficient facial landmark localization in visual images. hand_landmark. 2024/09/24: We release HaGRIDv2. 4 days ago · Hand landmark model bundle. 300 Videos in the Wild (300-VW) is a dataset for evaluating facial landmark tracking algorithms in the wild. Nov 15, 2021 · The updated version of our hand pose detection API improves the quality for 2D keypoint prediction, handedness (classification output whether it is left or right hand), and minimizes the number of false positive detections. The hand landmark tracking subgraph internally uses a hand landmark subgraph from the same module and a palm detection subgraph from the palm detection module. See full list on chuoling. The library supports dynamic hand gestures through trajectories and keyframe extraction. More details about the updated model can be found in our recent paper: On-device Real-time Hand Gesture Recognition. Carnegie Mellon University 4 days ago · These instructions show you how to use the Hand Landmarker with Python. Check out the MediaPipe documentation to learn more about configuration options that this solution supports. I believe they used a machine learning m Aug 19, 2019 · Our MediaPipe graph for hand tracking is shown below. We only evaluated the accuracy of our proposed model using a subset of 188 images from the larger dataset mentioned previously, which included 8 acupoints localization. You can use it for image classification or image detection tasks. Source: Large-scale Multiview 3D Hand Pose Dataset Jul 26, 2023 · In this story we are going to use MediaPipe’s Hands hand landmark model [4], as the dataset used only consists of images of hands. This paper presents Multi training. Kaggle uses cookies from Google to deliver and enhance the quality of its services and to analyze traffic. set the keypoints, number of class and the path of the dataset; load_data. We introduce the Google Landmarks Dataset v2 (GLDv2), a new benchmark for large-scale, fine-grained instance recognition and image retrieval in the Oct 9, 2024 · The dataset includes real images with the associated precise and automatic hand properties, such as landmark coordinates, velocities, orientations, and finger widths. A cleaned, large and robust dataset for ISL Gesture Classification Indian Sign Language Hand Landmarks Dataset | Kaggle Kaggle uses cookies from Google to deliver and enhance the quality of its services and to analyze traffic. Hand landmark localization is therefore useful in many diverse cases and scenarios. It's implemented via MediaPipe, a framework for building cross-platform ML solutions. py: Script used for training the CNN model on datasets of hand gestures or numbers. First, install the package:!pip install mediapipe-model-maker from mediapipe_model_maker import gesture_recognizer. Dec 15, 2021 · Hi, I am using mediapipe handsolution to create handlandmark dataset and I encountered some basic problem. A pretrained model is available as part of Google's MediaPipe framework. Here are the steps to run hand landmark detection using MediaPipe. The dataset was constructed from our large-scale SpeakingFaces dataset. Read more, Paper on arXiv. The training set contains samples from a single user only (Jonathan Tompson), while the test set contains samples from two users (Murphy Stein and Jonathan Tompson Face images and mark coordinates are required. A hierarchical extension of the dataset is presented in an under-submission Rendered Hand Pose (RHD) is a dataset for hand pose estimation. Its open-source framework enables the development of efficient and scalable deep learning solutions, making it highly adaptable to a variety of real-time applications [7] , [8] . For hand presence, we select a subset of real-world images as positive examples and sample on the region excluding annotated hand regions as negative examples. Hand Detection Within this research area, hand isolation and detection is an important component of sign language recognition. The landmark depth is represented by the z value pointed at the depth value of the wrist [5]. It was found that the library outperforms another publicly available HGR system - MediaPipe Solutions, on three diverse, real-world datasets. # center of the given ROI. The proposed model and pipeline architecture demonstrates real-time inference speed on mobile Apr 10, 2021 · After that we are going to take care of the hand landmarks detection. py --dataset dataset/ --color gray --set train python Mar 8, 2024 · The high interest for human hands derives from the crucial role they play in the majority of human activities. Sep 1, 2024 · MediaPipe is exceptional in its ability to provide accurate hand landmark tracking, which is critical for precise gesture recognition. ↳ 2 cells hidden Nov 29, 2019 · Hello I am currently looking for a dataset that would have the hand landmarks (the predicted 21 key points) as an input (data) and the hand poses (Yeah!, Rock'n'Roll!, Spreadetc) as the target. 0: 0. Only use this method when the HandLandmarker is created with running mode image . The 3D kinematic model of the hand provides 21 keypoints per hand: 4 keypoints per finger and one keypoint close to the wrist. github. Learn more It is an extension of COCO 2017 dataset with the same train/val split as COCO. Keywords: hand pose estimation, landmark localization, hand detection, @inproceedings {zimmermann2019freihand, title = {Freihand: A dataset for markerless capture of hand pose and shape from single rgb images}, author = {Zimmermann, Christian and Ceylan, Duygu and Yang, Jimei and Russell, Bryan and Argus, Max and Brox, Thomas}, booktitle = {Proceedings of the IEEE International Conference on Computer Vision}, pages = {813--822}, year = {2019}} points, we used 21 MediaPipe Hands 3D Landmark implemented on the 6 Hand Gesture Dataset. In this paper, we have presented a comparative analysis of visual emotion recognition when posture is taken into account. The pipeline is implemented as a MediaPipe graph that uses a hand landmark tracking subgraph from the hand landmark module, and renders using a dedicated hand renderer subgraph. js model on custom datasets or pre-existing ones (e. BlazePalm is a fast, light-weight 2-part hand landmark detector from Google Research. 3D Hand Pose is a multi-view hand pose dataset consisting of color images of hands and different kind of annotations for each: the bounding box and the 2D and 3D location on the joints in the hand. This phase ensures the accumulation of diverse and extensive datasets necessary for model training. , facial features, finger joints) from images or videos, enabling applications like gesture recognition, facial analysis, and augmented reality. This version of the dataset contains approximately 5 million images, split into 3 sets of images: train, index and test The InterHand2. In case you want to retrain the networks on new data you can adapt the code provided to your needs. Training The training took about 10 hours on a single NVIDIA 1080TI and was performed with the YOLOv3 default architecture. Sensor-based methods, such as those that use gloves or other wearable devices, have historically been more accurate, but Create custom hand landmark datasets with 60 features (20 landmarks / hand - X,Y and depth). The code sample described in these instructions is available on GitHub. You can get the index of all the individual landmarks using the below code: self-occluded regions are labeled in a “landmark-marched” style [59], i. Size: 120 images (12 subjects) Landmarks: 73 Free: Yes Example: Name: FEI Face Database Apr 3, 2023 · Sample images from the dataset Hand Landmark Extraction. 21 hand landmarks are detected and plotted on the input image. Here is an example of one annotated image. two models, a hand palm detector that provides an oriented bounding box of the hand and a hand landmark model that operates on the bounding box to obtain 2. testCNN. This version of the dataset contains approximately 5 million images, split into 3 sets of images: train, index and test. Mar 19, 2023 · Load the Hand Landmark model: The first step is to prepare the dataset by splitting it into training and testing sets. METHOD IMPLEMENTATION 3. 5, so we think this is the average parameter. We explore using subsets (e. python build_dlib_landmarks_xml. Sep 5, 2023 · Real-time hand segmentation is a key process in applications that require human–computer interaction, such as gesture recognition or augmented reality systems. landmark[0]. 38 The probability distribution of each landmark point is shown in Fig. Custom Dataset Creation: Capture hand gestures for each alphabet (A–Z) and save them as a structured dataset. 0 - 1. Now let‘s take a closer look at what the hand landmark model is actually predicting. detectForVideo(videoFrame, timestamp, imageProcessingOptions) Contribute to odil-T/Hand-Gesture-Recognition development by creating an account on GitHub. For hand landmark points, we used the keypoint localization of 21 3D hand-knuckle coordinates in the hand region, implemented by using MediaPipe. For example, human-labeled 2D facial landmark datasets fo- The ICVL dataset is a hand pose estimation dataset that consists of 330K training frames and 2 testing sequences with each 800 frames. See the definition of the 21 landmarks below: two publicly available datasets demonstrating the improved performance of the proposed pipeline compared to other state-of-the-art approaches. The Hailo Model Zoo includes pre-trained models and a full building and evaluation environment - hailo-ai/hailo_model_zoo Performs hand landmarks detection on the provided single image and waits synchronously for the response. Extensive experiments are performed to demonstrate the effectiveness of the proposed method. We do this with a call to the process method on our Hands object. json file and store into . This method receives as input a ndarray with an image in RGB and returns as output a NamedTuple object containing a collection of hand landmarks for the hands found in the image and a collection of handedness of the detected hands (if each hand is a left or Dec 2, 2023 · I am working on a project where I am trying to classify certain hand positions, recording hand landmark data via MediaPipe, and then using this information to train a model. May 31, 2022 · In the Tea Tasting Review dataset experiment, we used a Tea Review Experiment Dataset to validate the classification approach we chose as a classifier based on the human hand representation of This is the official repository of Decaf dataset: 3D deformations: per-vertex 3D deformations for FLAME head model. g. yaml. Dec 14, 2024 · Face and hand landmark detection using CNNs identifies key points (e. Figure 3: Architecture of our hand landmark model. GU2Net presents a novel idea to train the landmark detection network on multiple datasets at once. By using this model, a speech-impaired person can easily HAnd Gesture Recognition Image Dataset. The individual landmarks can be accessed by specifying the index of the required landmark. GU2Net achieves state-of-the-art performance for head and hand dataset (difficult to judge for chest dataset). 7. May 26, 2023 · This dataset is designed to train a ControlNet with human hands. Typically, image-based models need to be trained on . 8% AP among all known real-time object detectors with 30 FPS or higher on GPU V100. The landmarks are then classified into their respective gesture class. Facial landmark detection is a cornerstone in many facial analysis tasks such as face recognition, drowsiness detection, and facial expression recognition. One example of this is [15], which uses Google’s hand landmark model [22] to identify hand landmark coordinates, which serves as a second input channel to a CNN. Contacts: per-vertex contact signals for head and hand. These instructions will get you a copy of the project up and running on your local machine for development and testing purposes. com/mediapipe/solutions/vision/hand_landmarker ). t 2D landmark definitions. The dataset contains 2,556 thermal-visual image pairs of 142 subjects with manually annotated face bounding boxes and 54 facial landmarks. This extra class contains 123,589 samples. extract the labels from the . the landmark feature is arranged to 1D structured data for each image. 2D bounding box: bounding box for a head and for a hand. Our InterHand2. , landmark) computation. using six of the landmarks or only one) provided by finger-spelling based on the various hand gestures. py: Script for testing the performance of the trained CNN model on a test dataset. 6M dataset is the first large-scale real-captured dataset with accurate GT 3D interacting hand poses . Jan 1, 2022 · A 2D hand landmark dataset is generated based on the Youtube3D Hands image-sequence [28]. py: Integrates MediaPipe to perform real-time hand detection and display landmarks through the webcam. The skeleton dataset is created using the hand image dataset and is done to improve the accuracy of the model. blob: This is the model to classify the hand's gesture into ASL characters. output_stream Whose Hands Are These? Hand Detection and Hand-Body Association in the Wild. Learn more. Mar 22, 2024 · Researchers found better results working to detect palms rather than hands, and by using a regression model to locate hand landmarks. 🙏 . It includes hand landmarks detected by MediaPipe(for more information refer to: https://developers. X,Y and depth). Apr 3, 2020 · While image retrieval and instance recognition techniques are progressing rapidly, there is a need for challenging datasets to accurately measure their performance -- while posing novel challenges that are relevant for practical applications. Hand Landmark Detection dataset by NIT RAIPUR 4 days ago · In Video mode and Live stream mode, if the hand presence confidence score from the hand landmark model is below this threshold, Hand Landmarker triggers the palm detection model. left_hand_landmarks. The model has three outputs sharing a feature extractor. Hand Landmark Model After running palm detection over the whole image, our subsequent hand landmark model performs precise land- The project utilizes the MediaPipe library, which provides pre-trained machine learning models for various tasks, including hand landmark recognition. 4 days ago · This task operates on image data with a machine learning (ML) model as static data or a continuous stream and outputs hand landmarks in image coordinates, hand landmarks in world coordinates and handedness(left/right hand) of multiple detected hands. Two main approaches are: (1) using a large amount of photo data of hand gestures and (2) using the MediaPipe Hand landmark framework. Example: results. The resulting dataset will contain the extracted hand landmark positions from each image, rather than images themselves. The confidence parameter on media pipe hands is 0. For each frame, the RGBD data from 3 Kinects is provided: a frontal view and 2 side views. 3D landmark detection plays a pivotal role in various applications such as 3D registration, pose estimation, and virtual try-on. 2for more detail. Fast-Hand o ers high accuracy scores while reaching a speed of 25 frames per second on an NVIDIA Jetson TX2 graphics processing unit. It provides segmentation maps with 33 classes: three for each finger, palm, person, and background. The palm detector identifies the hand's position, while the hand landmark model pinpoints specific 2D coordinates of the hand. 5D landmarks. The pipeline consists of two models: 1) a palm detector, 2) a hand landmark model. I am having issues getting the configuration squared away compared to another model I built using image data. The hand landmark model detects and localizes 21 key points (landmarks) on the hand, including fingertips, joints, and the palm. points, we used 21 MediaPipe Hands 3D Landmark implemented on the 6 Hand Gesture Dataset. PSL 101 - Polish Sign Language (no website) Modelling and Recognition of Signed Expressions Using Subunits Obtained by Data–Driven Approach 2012, Oszust et al. ↳ 2 cells hidden Oct 15, 2023 · The german sign alphabet [1]. The NYU Hand pose dataset contains 8252 test-set and 72757 training-set frames of captured RGBD data with ground-truth hand-pose information. More specifically, three different skeletal data were considered: an adapted body pose from the predefined MediaPipe body pose landmark model, a predefined MediaPipe hand pose landmark model, and a combination of both. On the other hand, current 3D datasets leave much to be desired in terms of accuracy and consistency w. The dataset is collected from 10 different subjects with 16 hand joint annotations for each frame. The dataset authors collected a large number of long facial videos recorded in the wild. The key hand_landmark_6_shaves. Another example of this is [19], which uses skin mask- Nov 7, 2024 · Exemplary images from dataset with landmark-based model outputs depicting acupoint locations on the back side of the hand, including LI4, TE1, TE2, TE3, LI1, PC9, SI1, HT9. YOLOv7 surpasses all known object detectors in both speed and accuracy in the range from 5 FPS to 160 FPS and has the highest accuracy 56. Real-time Hand Landmark Detection: Leverages MediaPipe HandLandmarker for accurate hand landmark extraction. Sep 1, 2021 · Please provide details, for instance, if a method is novel, explain what aspect is novel and why this is interesting. e. The hand landmark model bundle detects the keypoint Dec 16, 2024 · Fig 1. The main challenge is Jan 10, 2023 · The holistic model produces 468 Face landmarks, 21 Left-Hand landmarks, and 21 Right-Hand landmarks. However, # LANDMARKS provides coordinates (in pixels) of a 3D object projected onto the # 2D image surface, while WORLD_LANDMARKS provides coordinates (in meters) of # the 3D object itself. AR/VR Controls: Improving user experience in augmented and virtual reality. 6M dataset is a large-scale real-captured dataset with accurate GT 3D interacting hand poses, used for 3D hand pose estimation The dataset contains 2. One key optimization MediaPipe provides is that the palm detector is only run as necessary (fairly infrequently), saving significant computation time. Oct 9, 2023 · Pre-processing of image dataset. In this user-independent model, CNN-based models are trained using a set of image and hand skeleton datasets. See Section2. The training set is used to build the random forest model, while the Image size normalizes to [0. , on the nearest visible boundary. The graph consists of two subgraphs—one for hand detection and one for hand keypoints (i. Now we can extract the Sep 27, 2024 · What applications can benefit from using the Hand Keypoints dataset? The Hand Keypoints dataset can be applied in various fields, including: Gesture Recognition: Enhancing human-computer interaction. The project processes input images or frames from videos/webcam Feb 18, 2023 · Overall, MediaPipe Hand Landmark Detection is a powerful tool for hand tracking and gesture recognition in a wide range of applications, and is likely to become even more important as technology Description: This note describes a data set consisting of 120 annotated monocular images of 12 different frontal human faces. hand_asl_6_shaves. The model was trained on approximately 30K real-world images, as well as several rendered synthetic hand models imposed over various backgrounds. Also, some images have no_gesture class if there is a second free hand in the frame. We provide scripts to train HandSegNet and PoseNet on the Rendered Hand Pose Dataset (RHD). 5 This paper presents Multi-view Leap2 Hand Pose Dataset (ML2HP Dataset), a new dataset for hand pose recognition, captured using a multi-view recording setup with two Leap Motion Controller 2 devices. 0. Converted using OpenVino's myriad_compiler. Train your own hand gesture recognition models A dataset of 11k hands and palm images for gender and age detection. 6M: A Dataset and Baseline for 3D Interacting Hand Pose Estimation from a Single RGB Image (ECCV 2020). Feb 10, 2022 · We are trying to achieve better accuracy using image and hand skeleton datasets, and this dataset is created using a separate program on the sign language image data set. Abstract. Main objective of this project is detecting the hand and calculating the ratios used for hand analysis automatically. mtpphpd qsypcxk lchdu ngtv quwvqh iryn jnh yfevl vwc lfwli
{"Title":"What is the best girl name?","Description":"Wheel of girl names","FontSize":7,"LabelsList":["Emma","Olivia","Isabel","Sophie","Charlotte","Mia","Amelia","Harper","Evelyn","Abigail","Emily","Elizabeth","Mila","Ella","Avery","Camilla","Aria","Scarlett","Victoria","Madison","Luna","Grace","Chloe","Penelope","Riley","Zoey","Nora","Lily","Eleanor","Hannah","Lillian","Addison","Aubrey","Ellie","Stella","Natalia","Zoe","Leah","Hazel","Aurora","Savannah","Brooklyn","Bella","Claire","Skylar","Lucy","Paisley","Everly","Anna","Caroline","Nova","Genesis","Emelia","Kennedy","Maya","Willow","Kinsley","Naomi","Sarah","Allison","Gabriella","Madelyn","Cora","Eva","Serenity","Autumn","Hailey","Gianna","Valentina","Eliana","Quinn","Nevaeh","Sadie","Linda","Alexa","Josephine","Emery","Julia","Delilah","Arianna","Vivian","Kaylee","Sophie","Brielle","Madeline","Hadley","Ibby","Sam","Madie","Maria","Amanda","Ayaana","Rachel","Ashley","Alyssa","Keara","Rihanna","Brianna","Kassandra","Laura","Summer","Chelsea","Megan","Jordan"],"Style":{"_id":null,"Type":0,"Colors":["#f44336","#710d06","#9c27b0","#3e1046","#03a9f4","#014462","#009688","#003c36","#8bc34a","#38511b","#ffeb3b","#7e7100","#ff9800","#663d00","#607d8b","#263238","#e91e63","#600927","#673ab7","#291749","#2196f3","#063d69","#00bcd4","#004b55","#4caf50","#1e4620","#cddc39","#575e11","#ffc107","#694f00","#9e9e9e","#3f3f3f","#3f51b5","#192048","#ff5722","#741c00","#795548","#30221d"],"Data":[[0,1],[2,3],[4,5],[6,7],[8,9],[10,11],[12,13],[14,15],[16,17],[18,19],[20,21],[22,23],[24,25],[26,27],[28,29],[30,31],[0,1],[2,3],[32,33],[4,5],[6,7],[8,9],[10,11],[12,13],[14,15],[16,17],[18,19],[20,21],[22,23],[24,25],[26,27],[28,29],[34,35],[30,31],[0,1],[2,3],[32,33],[4,5],[6,7],[10,11],[12,13],[14,15],[16,17],[18,19],[20,21],[22,23],[24,25],[26,27],[28,29],[34,35],[30,31],[0,1],[2,3],[32,33],[6,7],[8,9],[10,11],[12,13],[16,17],[20,21],[22,23],[26,27],[28,29],[30,31],[0,1],[2,3],[32,33],[4,5],[6,7],[8,9],[10,11],[12,13],[14,15],[18,19],[20,21],[22,23],[24,25],[26,27],[28,29],[34,35],[30,31],[0,1],[2,3],[32,33],[4,5],[6,7],[8,9],[10,11],[12,13],[36,37],[14,15],[16,17],[18,19],[20,21],[22,23],[24,25],[26,27],[28,29],[34,35],[30,31],[2,3],[32,33],[4,5],[6,7]],"Space":null},"ColorLock":null,"LabelRepeat":1,"ThumbnailUrl":"","Confirmed":true,"TextDisplayType":null,"Flagged":false,"DateModified":"2020-02-05T05:14:","CategoryId":3,"Weights":[],"WheelKey":"what-is-the-best-girl-name"}