├── .gitignore ├── CODE_OF_CONDUCT.md ├── LICENSE ├── README.md ├── SECURITY.md ├── SUPPORT.md ├── docs ├── _config.yml ├── css │ └── styles.css ├── img │ ├── Microsoft-logo-only.svg │ ├── Microsoft-logo.svg │ ├── Microsoft-logo_rgb_c-gray.png │ ├── Microsoft-logo_rgb_c-wht.png │ ├── Microsoft-logo_rgb_wht.png │ ├── dataset_samples.jpg │ ├── favicon.ico │ ├── joints_uncertainty.png │ ├── plot_20_L_Wrist.png │ ├── plot_21_R_Wrist.png │ ├── plot_all_bold.png │ └── results.png └── index.html └── download_data.py /.gitignore: -------------------------------------------------------------------------------- 1 | ## Ignore Visual Studio temporary files, build results, and 2 | ## files generated by popular Visual Studio add-ons. 3 | ## 4 | ## Get latest from https://github.com/github/gitignore/blob/main/VisualStudio.gitignore 5 | 6 | # User-specific files 7 | *.rsuser 8 | *.suo 9 | *.user 10 | *.userosscache 11 | *.sln.docstates 12 | 13 | # User-specific files (MonoDevelop/Xamarin Studio) 14 | *.userprefs 15 | 16 | # Mono auto generated files 17 | mono_crash.* 18 | 19 | # Build results 20 | [Dd]ebug/ 21 | [Dd]ebugPublic/ 22 | [Rr]elease/ 23 | [Rr]eleases/ 24 | x64/ 25 | x86/ 26 | [Ww][Ii][Nn]32/ 27 | [Aa][Rr][Mm]/ 28 | [Aa][Rr][Mm]64/ 29 | bld/ 30 | [Bb]in/ 31 | [Oo]bj/ 32 | [Ll]og/ 33 | [Ll]ogs/ 34 | 35 | # Visual Studio 2015/2017 cache/options directory 36 | .vs/ 37 | # Uncomment if you have tasks that create the project's static files in wwwroot 38 | #wwwroot/ 39 | 40 | # Visual Studio 2017 auto generated files 41 | Generated\ Files/ 42 | 43 | # MSTest test Results 44 | [Tt]est[Rr]esult*/ 45 | [Bb]uild[Ll]og.* 46 | 47 | # NUnit 48 | *.VisualState.xml 49 | TestResult.xml 50 | nunit-*.xml 51 | 52 | # Build Results of an ATL Project 53 | [Dd]ebugPS/ 54 | [Rr]eleasePS/ 55 | dlldata.c 56 | 57 | # Benchmark Results 58 | BenchmarkDotNet.Artifacts/ 59 | 60 | # .NET Core 61 | project.lock.json 62 | project.fragment.lock.json 63 | artifacts/ 64 | 65 | # ASP.NET Scaffolding 66 | ScaffoldingReadMe.txt 67 | 68 | # StyleCop 69 | StyleCopReport.xml 70 | 71 | # Files built by Visual Studio 72 | *_i.c 73 | *_p.c 74 | *_h.h 75 | *.ilk 76 | *.meta 77 | *.obj 78 | *.iobj 79 | *.pch 80 | *.pdb 81 | *.ipdb 82 | *.pgc 83 | *.pgd 84 | *.rsp 85 | *.sbr 86 | *.tlb 87 | *.tli 88 | *.tlh 89 | *.tmp 90 | *.tmp_proj 91 | *_wpftmp.csproj 92 | *.log 93 | *.tlog 94 | *.vspscc 95 | *.vssscc 96 | .builds 97 | *.pidb 98 | *.svclog 99 | *.scc 100 | 101 | # Chutzpah Test files 102 | _Chutzpah* 103 | 104 | # Visual C++ cache files 105 | ipch/ 106 | *.aps 107 | *.ncb 108 | *.opendb 109 | *.opensdf 110 | *.sdf 111 | *.cachefile 112 | *.VC.db 113 | *.VC.VC.opendb 114 | 115 | # Visual Studio profiler 116 | *.psess 117 | *.vsp 118 | *.vspx 119 | *.sap 120 | 121 | # Visual Studio Trace Files 122 | *.e2e 123 | 124 | # TFS 2012 Local Workspace 125 | $tf/ 126 | 127 | # Guidance Automation Toolkit 128 | *.gpState 129 | 130 | # ReSharper is a .NET coding add-in 131 | _ReSharper*/ 132 | *.[Rr]e[Ss]harper 133 | *.DotSettings.user 134 | 135 | # TeamCity is a build add-in 136 | _TeamCity* 137 | 138 | # DotCover is a Code Coverage Tool 139 | *.dotCover 140 | 141 | # AxoCover is a Code Coverage Tool 142 | .axoCover/* 143 | !.axoCover/settings.json 144 | 145 | # Coverlet is a free, cross platform Code Coverage Tool 146 | coverage*.json 147 | coverage*.xml 148 | coverage*.info 149 | 150 | # Visual Studio code coverage results 151 | *.coverage 152 | *.coveragexml 153 | 154 | # NCrunch 155 | _NCrunch_* 156 | .*crunch*.local.xml 157 | nCrunchTemp_* 158 | 159 | # MightyMoose 160 | *.mm.* 161 | AutoTest.Net/ 162 | 163 | # Web workbench (sass) 164 | .sass-cache/ 165 | 166 | # Installshield output folder 167 | [Ee]xpress/ 168 | 169 | # DocProject is a documentation generator add-in 170 | DocProject/buildhelp/ 171 | DocProject/Help/*.HxT 172 | DocProject/Help/*.HxC 173 | DocProject/Help/*.hhc 174 | DocProject/Help/*.hhk 175 | DocProject/Help/*.hhp 176 | DocProject/Help/Html2 177 | DocProject/Help/html 178 | 179 | # Click-Once directory 180 | publish/ 181 | 182 | # Publish Web Output 183 | *.[Pp]ublish.xml 184 | *.azurePubxml 185 | # Note: Comment the next line if you want to checkin your web deploy settings, 186 | # but database connection strings (with potential passwords) will be unencrypted 187 | *.pubxml 188 | *.publishproj 189 | 190 | # Microsoft Azure Web App publish settings. Comment the next line if you want to 191 | # checkin your Azure Web App publish settings, but sensitive information contained 192 | # in these scripts will be unencrypted 193 | PublishScripts/ 194 | 195 | # NuGet Packages 196 | *.nupkg 197 | # NuGet Symbol Packages 198 | *.snupkg 199 | # The packages folder can be ignored because of Package Restore 200 | **/[Pp]ackages/* 201 | # except build/, which is used as an MSBuild target. 202 | !**/[Pp]ackages/build/ 203 | # Uncomment if necessary however generally it will be regenerated when needed 204 | #!**/[Pp]ackages/repositories.config 205 | # NuGet v3's project.json files produces more ignorable files 206 | *.nuget.props 207 | *.nuget.targets 208 | 209 | # Microsoft Azure Build Output 210 | csx/ 211 | *.build.csdef 212 | 213 | # Microsoft Azure Emulator 214 | ecf/ 215 | rcf/ 216 | 217 | # Windows Store app package directories and files 218 | AppPackages/ 219 | BundleArtifacts/ 220 | Package.StoreAssociation.xml 221 | _pkginfo.txt 222 | *.appx 223 | *.appxbundle 224 | *.appxupload 225 | 226 | # Visual Studio cache files 227 | # files ending in .cache can be ignored 228 | *.[Cc]ache 229 | # but keep track of directories ending in .cache 230 | !?*.[Cc]ache/ 231 | 232 | # Others 233 | ClientBin/ 234 | ~$* 235 | *~ 236 | *.dbmdl 237 | *.dbproj.schemaview 238 | *.jfm 239 | *.pfx 240 | *.publishsettings 241 | orleans.codegen.cs 242 | 243 | # Including strong name files can present a security risk 244 | # (https://github.com/github/gitignore/pull/2483#issue-259490424) 245 | #*.snk 246 | 247 | # Since there are multiple workflows, uncomment next line to ignore bower_components 248 | # (https://github.com/github/gitignore/pull/1529#issuecomment-104372622) 249 | #bower_components/ 250 | 251 | # RIA/Silverlight projects 252 | Generated_Code/ 253 | 254 | # Backup & report files from converting an old project file 255 | # to a newer Visual Studio version. Backup files are not needed, 256 | # because we have git ;-) 257 | _UpgradeReport_Files/ 258 | Backup*/ 259 | UpgradeLog*.XML 260 | UpgradeLog*.htm 261 | ServiceFabricBackup/ 262 | *.rptproj.bak 263 | 264 | # SQL Server files 265 | *.mdf 266 | *.ldf 267 | *.ndf 268 | 269 | # Business Intelligence projects 270 | *.rdl.data 271 | *.bim.layout 272 | *.bim_*.settings 273 | *.rptproj.rsuser 274 | *- [Bb]ackup.rdl 275 | *- [Bb]ackup ([0-9]).rdl 276 | *- [Bb]ackup ([0-9][0-9]).rdl 277 | 278 | # Microsoft Fakes 279 | FakesAssemblies/ 280 | 281 | # GhostDoc plugin setting file 282 | *.GhostDoc.xml 283 | 284 | # Node.js Tools for Visual Studio 285 | .ntvs_analysis.dat 286 | node_modules/ 287 | 288 | # Visual Studio 6 build log 289 | *.plg 290 | 291 | # Visual Studio 6 workspace options file 292 | *.opt 293 | 294 | # Visual Studio 6 auto-generated workspace file (contains which files were open etc.) 295 | *.vbw 296 | 297 | # Visual Studio 6 auto-generated project file (contains which files were open etc.) 298 | *.vbp 299 | 300 | # Visual Studio 6 workspace and project file (working project files containing files to include in project) 301 | *.dsw 302 | *.dsp 303 | 304 | # Visual Studio 6 technical files 305 | *.ncb 306 | *.aps 307 | 308 | # Visual Studio LightSwitch build output 309 | **/*.HTMLClient/GeneratedArtifacts 310 | **/*.DesktopClient/GeneratedArtifacts 311 | **/*.DesktopClient/ModelManifest.xml 312 | **/*.Server/GeneratedArtifacts 313 | **/*.Server/ModelManifest.xml 314 | _Pvt_Extensions 315 | 316 | # Paket dependency manager 317 | .paket/paket.exe 318 | paket-files/ 319 | 320 | # FAKE - F# Make 321 | .fake/ 322 | 323 | # CodeRush personal settings 324 | .cr/personal 325 | 326 | # Python Tools for Visual Studio (PTVS) 327 | __pycache__/ 328 | *.pyc 329 | 330 | # Cake - Uncomment if you are using it 331 | # tools/** 332 | # !tools/packages.config 333 | 334 | # Tabs Studio 335 | *.tss 336 | 337 | # Telerik's JustMock configuration file 338 | *.jmconfig 339 | 340 | # BizTalk build output 341 | *.btp.cs 342 | *.btm.cs 343 | *.odx.cs 344 | *.xsd.cs 345 | 346 | # OpenCover UI analysis results 347 | OpenCover/ 348 | 349 | # Azure Stream Analytics local run output 350 | ASALocalRun/ 351 | 352 | # MSBuild Binary and Structured Log 353 | *.binlog 354 | 355 | # NVidia Nsight GPU debugger configuration file 356 | *.nvuser 357 | 358 | # MFractors (Xamarin productivity tool) working folder 359 | .mfractor/ 360 | 361 | # Local History for Visual Studio 362 | .localhistory/ 363 | 364 | # Visual Studio History (VSHistory) files 365 | .vshistory/ 366 | 367 | # BeatPulse healthcheck temp database 368 | healthchecksdb 369 | 370 | # Backup folder for Package Reference Convert tool in Visual Studio 2017 371 | MigrationBackup/ 372 | 373 | # Ionide (cross platform F# VS Code tools) working folder 374 | .ionide/ 375 | 376 | # Fody - auto-generated XML schema 377 | FodyWeavers.xsd 378 | 379 | # VS Code files for those working on multiple tools 380 | .vscode/* 381 | !.vscode/settings.json 382 | !.vscode/tasks.json 383 | !.vscode/launch.json 384 | !.vscode/extensions.json 385 | *.code-workspace 386 | 387 | # Local History for Visual Studio Code 388 | .history/ 389 | 390 | # Windows Installer files from build outputs 391 | *.cab 392 | *.msi 393 | *.msix 394 | *.msm 395 | *.msp 396 | 397 | # JetBrains Rider 398 | *.sln.iml 399 | -------------------------------------------------------------------------------- /CODE_OF_CONDUCT.md: -------------------------------------------------------------------------------- 1 | # Microsoft Open Source Code of Conduct 2 | 3 | This project has adopted the [Microsoft Open Source Code of Conduct](https://opensource.microsoft.com/codeofconduct/). 4 | 5 | Resources: 6 | 7 | - [Microsoft Open Source Code of Conduct](https://opensource.microsoft.com/codeofconduct/) 8 | - [Microsoft Code of Conduct FAQ](https://opensource.microsoft.com/codeofconduct/faq/) 9 | - Contact [opencode@microsoft.com](mailto:opencode@microsoft.com) with questions or concerns 10 | -------------------------------------------------------------------------------- /LICENSE: -------------------------------------------------------------------------------- 1 | Research Use of Data Agreement v1.0 2 | 3 | This is the Research Use of Data Agreement, Version 1.0 (the “R-UDA”). Capitalized terms are defined in Section 5. Data Provider and you agree as follows: 4 | 5 | 1. Provision of the Data 6 | 1.1. You may use, modify, and distribute the Data made available to you by the Data Provider under this R-UDA for Research Use if you follow the R-UDA’s terms. 7 | 1.2. Data Provider will not sue you or any Downstream Recipient for any claim arising out of the use, modification, or distribution of the Data provided you meet the terms of the R-UDA. 8 | 1.3. This R-UDA does not restrict your use, modification, or distribution of any portions of the Data that are in the public domain or that may be used, modified, or distributed under any other legal exception or limitation. 9 | 10 | 2. Restrictions 11 | 2.1. You agree that you will use the Data solely for Computational Use for non-commercial research. This restriction means that you may engage in non-commercial research activities (including non-commercial research undertaken by or funded via a commercial entity), but you may not use the Data or any Results in any commercial offering, including as part of a product or service (or to improve any product or service) you use or provide to others. 12 | 2.2. You may not receive money or other consideration in exchange for use or redistribution of Data. 13 | 14 | 3. Redistribution of Data 15 | 3.1. You may redistribute the Data, so long as: 16 | 3.1.1. You include with any Data you redistribute all credit or attribution information that you received with the Data, and your terms require any Downstream Recipient to do the same; and 17 | 3.1.2. You bind each recipient to whom you redistribute the Data to the terms of the R-UDA. 18 | 19 | 4. No Warranty, Limitation of Liability 20 | 4.1. Data Provider does not represent or warrant that it has any rights whatsoever in the Data. 21 | 4.2. THE DATA IS PROVIDED ON AN “AS IS” BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, EITHER EXPRESS OR IMPLIED INCLUDING, WITHOUT LIMITATION, ANY WARRANTIES OR CONDITIONS OF TITLE, NON-INFRINGEMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. 22 | 4.3. NEITHER DATA PROVIDER NOR ANY UPSTREAM DATA PROVIDER SHALL HAVE ANY LIABILITY FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING WITHOUT LIMITATION LOST PROFITS), HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE DATA OR RESULTS, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGES. 23 | 24 | 5. Definitions 25 | 5.1. “Computational Use” means activities necessary to enable the use of Data (alone or along with other material) for analysis by a computer. 26 | 5.2. “Data” means the material you receive under the R-UDA in modified or unmodified form, but not including Results. 27 | 5.3. “Data Provider” means the source from which you receive the Data and with whom you enter into the R-UDA. 28 | 5.4. “Downstream Recipient” means any person or persons who receives the Data directly or indirectly from you in accordance with the R-UDA. 29 | 5.5. “Result” means anything that you develop or improve from your use of Data that does not include more than a de minimis portion of the Data on which the use is based. Results may include de minimis portions of the Data necessary to report on or explain use that has been conducted with the Data, such as figures in scientific papers, but do not include more. Artificial intelligence models trained on Data (and which do not include more than a de minimis portion of Data) are Results. 30 | 5.6. “Upstream Data Providers” means the source or sources from which the Data Provider directly or indirectly received, under the terms of the R-UDA, material that is included in the Data. 31 | -------------------------------------------------------------------------------- /README.md: -------------------------------------------------------------------------------- 1 |  2 | 3 | # The SynthEgo Dataset 4 | 5 | The SynthEgo dataset was introduced in our paper [**SimpleEgo: Predicting probabilistic body pose from egocentric cameras**](https://microsoft.github.io/SimpleEgo/). 6 | 7 | The dataset contains: 8 | 9 | - 60,000 stereo pair synthetic RGB images from a head mounted camera at 1280x720 pixel resolution. 10 | - [SMPL-H](https://mano.is.tue.mpg.de/) pose and identity parameters for each stereo pair. 11 | - 3D joint locations in world and camera space for each image. 12 | - 2D joint locations in image space for each image. 13 | - Camera parameters for each image. 14 | 15 | ## Downloading the dataset 16 | 17 | The license terms for the [MANO](https://mano.is.tue.mpg.de/) dataset and parts of the [AMASS](https://amass.is.tue.mpg.de/) dataset prevent redistribution. 18 | As such, we include names and indices for these poses in the dataset we distribute, but not the pose parameters themselves. 19 | **For simplicity, we provide a script to download the MANO, AMASS and SynthEgo datasets and splice in the relevant pose parameters automatically.** 20 | You will need to sign up for MANO and AMASS and provide the relevant credentials when prompted by the script. 21 | The only requirements for the script are numpy and wget, otherwise simply run `python download_dataset.py` from the command line. 22 | 23 | You can download the dataset in parts (3GB each) directly from the following links, though *this is not recommended*: 24 | 25 | - [Part 1](https://facesyntheticspubwedata.z6.web.core.windows.net/3dv-2024/synth_ego_01.zip) 26 | - [Part 2](https://facesyntheticspubwedata.z6.web.core.windows.net/3dv-2024/synth_ego_02.zip) 27 | - [Part 3](https://facesyntheticspubwedata.z6.web.core.windows.net/3dv-2024/synth_ego_03.zip) 28 | - [Part 4](https://facesyntheticspubwedata.z6.web.core.windows.net/3dv-2024/synth_ego_04.zip) 29 | - [Part 5](https://facesyntheticspubwedata.z6.web.core.windows.net/3dv-2024/synth_ego_05.zip) 30 | - [Part 6](https://facesyntheticspubwedata.z6.web.core.windows.net/3dv-2024/synth_ego_06.zip) 31 | - [Part 7](https://facesyntheticspubwedata.z6.web.core.windows.net/3dv-2024/synth_ego_07.zip) 32 | - [Part 7](https://facesyntheticspubwedata.z6.web.core.windows.net/3dv-2024/synth_ego_08.zip) 33 | - [Part 9](https://facesyntheticspubwedata.z6.web.core.windows.net/3dv-2024/synth_ego_09.zip) 34 | - [Part 10](https://facesyntheticspubwedata.z6.web.core.windows.net/3dv-2024/synth_ego_10.zip) 35 | 36 | ## Dataset layout 37 | 38 | Once extracted, the dataset contains files for each sample of the form: 39 | 40 | - `metadata_0000000_0000.json` 41 | - `img_L_0000000_0000.jpg` 42 | - `img_R_0000000_0000.jpg` 43 | 44 | Where the first number indicates the subject index from 0 to 6000, and the second number indicates the frame index from 0 to 10. 45 | The metadata files are structured as follows: 46 | 47 | ```json 48 | { 49 | "pose": [ "52x3 array of SMPL-H thetas/pose parameters" ], 50 | "translation": [ "3 element SMPL-H translation vector" ], 51 | "identity": [ "10 element SMPL-H neutral beta vector" ], 52 | "cameras": { 53 | "camera_L": { 54 | "world_to_camera": [ "4x4 extrinsic matrix" ], 55 | "camera_to_image": [ "3x3 intrinsic matrix" ], 56 | "resolution": [1280, 720] 57 | }, 58 | "camera_R": { "as for camera_L" }, 59 | }, 60 | "landmarks": { 61 | "3D_world": [ "54x3 joint locations in world space" ], 62 | "3D_camera_L": [ "54x3 joint locations in camera space for camera_L" ], 63 | "2D_camera_L": [ "54x2 joint locations in image space for camera_L" ], 64 | "3D_camera_R": [ "54x3 joint locations in camera space for camera_R" ], 65 | "2D_camera_R": [ "54x2 joint locations in image space for camera_R" ], 66 | } 67 | } 68 | ``` 69 | 70 | ## Citation 71 | 72 | If you use the SynthEgo Dataset your research, please cite the following paper: 73 | 74 | ```bibtex 75 | @inproceedings{cuevas2024simpleego, 76 | title={{SimpleEgo}: Predicting probabilistic body pose from egocentric cameras}, 77 | author={Cuevas-Velasquez, Hanz and Hewitt, Charlie and Aliakbarian, Sadegh and Baltru{\v{s}}aitis, Tadas}, 78 | booktitle={2024 International Conference on 3D Vision (3DV)}, 79 | pages={1446--1455}, 80 | year={2024}, 81 | organization={IEEE} 82 | } 83 | ``` 84 | -------------------------------------------------------------------------------- /SECURITY.md: -------------------------------------------------------------------------------- 1 | 2 | 3 | ## Security 4 | 5 | Microsoft takes the security of our software products and services seriously, which includes all source code repositories managed through our GitHub organizations, which include [Microsoft](https://github.com/Microsoft), [Azure](https://github.com/Azure), [DotNet](https://github.com/dotnet), [AspNet](https://github.com/aspnet) and [Xamarin](https://github.com/xamarin). 6 | 7 | If you believe you have found a security vulnerability in any Microsoft-owned repository that meets [Microsoft's definition of a security vulnerability](https://aka.ms/security.md/definition), please report it to us as described below. 8 | 9 | ## Reporting Security Issues 10 | 11 | **Please do not report security vulnerabilities through public GitHub issues.** 12 | 13 | Instead, please report them to the Microsoft Security Response Center (MSRC) at [https://msrc.microsoft.com/create-report](https://aka.ms/security.md/msrc/create-report). 14 | 15 | If you prefer to submit without logging in, send email to [secure@microsoft.com](mailto:secure@microsoft.com). If possible, encrypt your message with our PGP key; please download it from the [Microsoft Security Response Center PGP Key page](https://aka.ms/security.md/msrc/pgp). 16 | 17 | You should receive a response within 24 hours. If for some reason you do not, please follow up via email to ensure we received your original message. Additional information can be found at [microsoft.com/msrc](https://www.microsoft.com/msrc). 18 | 19 | Please include the requested information listed below (as much as you can provide) to help us better understand the nature and scope of the possible issue: 20 | 21 | * Type of issue (e.g. buffer overflow, SQL injection, cross-site scripting, etc.) 22 | * Full paths of source file(s) related to the manifestation of the issue 23 | * The location of the affected source code (tag/branch/commit or direct URL) 24 | * Any special configuration required to reproduce the issue 25 | * Step-by-step instructions to reproduce the issue 26 | * Proof-of-concept or exploit code (if possible) 27 | * Impact of the issue, including how an attacker might exploit the issue 28 | 29 | This information will help us triage your report more quickly. 30 | 31 | If you are reporting for a bug bounty, more complete reports can contribute to a higher bounty award. Please visit our [Microsoft Bug Bounty Program](https://aka.ms/security.md/msrc/bounty) page for more details about our active programs. 32 | 33 | ## Preferred Languages 34 | 35 | We prefer all communications to be in English. 36 | 37 | ## Policy 38 | 39 | Microsoft follows the principle of [Coordinated Vulnerability Disclosure](https://aka.ms/security.md/cvd). 40 | 41 | 42 | -------------------------------------------------------------------------------- /SUPPORT.md: -------------------------------------------------------------------------------- 1 | # TODO: The maintainer of this repo has not yet edited this file 2 | 3 | **REPO OWNER**: Do you want Customer Service & Support (CSS) support for this product/project? 4 | 5 | - **No CSS support:** Fill out this template with information about how to file issues and get help. 6 | - **Yes CSS support:** Fill out an intake form at [aka.ms/onboardsupport](https://aka.ms/onboardsupport). CSS will work with/help you to determine next steps. 7 | - **Not sure?** Fill out an intake as though the answer were "Yes". CSS will help you decide. 8 | 9 | *Then remove this first heading from this SUPPORT.MD file before publishing your repo.* 10 | 11 | # Support 12 | 13 | ## How to file issues and get help 14 | 15 | This project uses GitHub Issues to track bugs and feature requests. Please search the existing 16 | issues before filing new issues to avoid duplicates. For new issues, file your bug or 17 | feature request as a new Issue. 18 | 19 | For help and questions about using this project, please **REPO MAINTAINER: INSERT INSTRUCTIONS HERE 20 | FOR HOW TO ENGAGE REPO OWNERS OR COMMUNITY FOR HELP. COULD BE A STACK OVERFLOW TAG OR OTHER 21 | CHANNEL. WHERE WILL YOU HELP PEOPLE?**. 22 | 23 | ## Microsoft Support Policy 24 | 25 | Support for this **PROJECT or PRODUCT** is limited to the resources listed above. 26 | -------------------------------------------------------------------------------- /docs/_config.yml: -------------------------------------------------------------------------------- 1 | theme: jekyll-theme-cayman -------------------------------------------------------------------------------- /docs/css/styles.css: -------------------------------------------------------------------------------- 1 | @import "https://cdn.jsdelivr.net/npm/bulma@0.9.3/css/bulma.min.css"; 2 | @import "https://cdn.jsdelivr.net/npm/@creativebulma/bulma-tooltip@1.2.0/dist/bulma-tooltip.min.css"; 3 | @media screen and (min-width: 1024px) { 4 | .container { 5 | max-width: 860px; 6 | } 7 | } 8 | 9 | .authors>span { 10 | padding: 0 0.5rem; 11 | display: inline-block; 12 | } 13 | 14 | @media only screen and (max-width: 480px) { 15 | a.button.is-rounded.is-link.is-light:not(:last-child) { 16 | margin-bottom: 0.75em; 17 | } 18 | } 19 | 20 | .footer-links a { 21 | color: inherit !important; 22 | padding: 0 1em; 23 | } -------------------------------------------------------------------------------- /docs/img/Microsoft-logo-only.svg: -------------------------------------------------------------------------------- 1 | 2 | 3 | 20 | -------------------------------------------------------------------------------- /docs/img/Microsoft-logo.svg: -------------------------------------------------------------------------------- 1 | 2 | 3 | 46 | -------------------------------------------------------------------------------- /docs/img/Microsoft-logo_rgb_c-gray.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/microsoft/SimpleEgo/c5b171e5aaabad62ee1f22dbe81f0c9de49168d9/docs/img/Microsoft-logo_rgb_c-gray.png -------------------------------------------------------------------------------- /docs/img/Microsoft-logo_rgb_c-wht.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/microsoft/SimpleEgo/c5b171e5aaabad62ee1f22dbe81f0c9de49168d9/docs/img/Microsoft-logo_rgb_c-wht.png -------------------------------------------------------------------------------- /docs/img/Microsoft-logo_rgb_wht.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/microsoft/SimpleEgo/c5b171e5aaabad62ee1f22dbe81f0c9de49168d9/docs/img/Microsoft-logo_rgb_wht.png -------------------------------------------------------------------------------- /docs/img/dataset_samples.jpg: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/microsoft/SimpleEgo/c5b171e5aaabad62ee1f22dbe81f0c9de49168d9/docs/img/dataset_samples.jpg -------------------------------------------------------------------------------- /docs/img/favicon.ico: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/microsoft/SimpleEgo/c5b171e5aaabad62ee1f22dbe81f0c9de49168d9/docs/img/favicon.ico -------------------------------------------------------------------------------- /docs/img/joints_uncertainty.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/microsoft/SimpleEgo/c5b171e5aaabad62ee1f22dbe81f0c9de49168d9/docs/img/joints_uncertainty.png -------------------------------------------------------------------------------- /docs/img/plot_20_L_Wrist.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/microsoft/SimpleEgo/c5b171e5aaabad62ee1f22dbe81f0c9de49168d9/docs/img/plot_20_L_Wrist.png -------------------------------------------------------------------------------- /docs/img/plot_21_R_Wrist.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/microsoft/SimpleEgo/c5b171e5aaabad62ee1f22dbe81f0c9de49168d9/docs/img/plot_21_R_Wrist.png -------------------------------------------------------------------------------- /docs/img/plot_all_bold.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/microsoft/SimpleEgo/c5b171e5aaabad62ee1f22dbe81f0c9de49168d9/docs/img/plot_all_bold.png -------------------------------------------------------------------------------- /docs/img/results.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/microsoft/SimpleEgo/c5b171e5aaabad62ee1f22dbe81f0c9de49168d9/docs/img/results.png -------------------------------------------------------------------------------- /docs/index.html: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 |
5 | 6 | 7 |63 | Predicting probabilistic body pose from egocentric cameras 64 |
65 |66 | International Conference on 3D Vision 2024 67 |
68 | 69 | 83 |113 | Our work addresses the problem of egocentric human pose estimation from downwards-facing cameras on 114 | head-mounted devices (HMD). 115 | This presents a challenging scenario, as parts of the body often fall outside of the image or are 116 | occluded. 117 | Previous solutions minimize this problem by using fish-eye camera lenses to capture a wider view, 118 | but these can present hardware design issues. 119 | They also predict 2D heat-maps per joint and lift them to 3D space to deal with self-occlusions, but 120 | this requires large network architectures which are impractical to deploy on resource-constrained 121 | HMDs. 122 | We predict pose from images captured with conventional rectilinear camera lenses. 123 | This resolves hardware design issues, but means body parts are often out of frame. 124 | As such, we directly regress probabilistic joint rotations represented as matrix Fisher 125 | distributions for a parameterized body model. 126 | This allows us to quantify pose uncertainties and explain out-of-frame or occluded joints. 127 | This also removes the need to compute 2D heat-maps and allows for simplified DNN architectures which 128 | require less compute. 129 | Given the lack of egocentric datasets using rectilinear camera lenses, we introduce the 130 | SynthEgo dataset, a synthetic dataset with 60K stereo images containing high diversity of 131 | pose, shape, clothing and skin tone. 132 | Our approach achieves state-of-the-art results for this challenging configuration, reducing mean 133 | per-joint position error by 23% overall and 58% for the lower body. 134 | Our architecture also has eight times fewer parameters and runs twice as fast as the current 135 | state-of-the-art. 136 | Experiments show that training on our synthetic dataset leads to good generalization to real world 137 | images without fine-tuning. 138 |
139 |152 | To construct the SynthEgo dataset we render 60K stereo pairs at 1280×720 pixel 153 | resolution, building on the pipeline of Hewitt et al. 154 | This dataset is comprised of 6000 unique identities, each performing 10 different poses in 10 155 | different lighting environments. 156 | Each identity is made up of a randomly sampled body shape, skin textures sampled from a library of 157 | 25 and randomly recolored, and clothing assets sampled from a library of 202. 158 | Lighting environments are sampled from a library of 489 HDRIs, to ensure correct disparity of the 159 | environment between the stereo pair, we project the HDRI background onto the ground plane. 160 | Poses are sampled from a library of over 2 million unique poses and randomly mirrored; sampling is 161 | weighted by the mean absolute joint angle and common poses like T-pose are significantly 162 | down-weighted to increase diversity. 163 |
164 | 165 |168 | | Mo2Cap2 170 | | 171 |xR-EgoPose | 172 |UnrealEgo | 173 |SynthEgo | 174 | 175 | 176 |
---|---|---|---|---|
Unique Identities | 178 |700 | 179 |46 | 180 |17 | 181 |6000 | 182 |
Environments | 185 |Unspecified | 186 |Unspecified | 187 |14 | 188 |489 | 189 |
Body Model | 192 |SMPL | 193 |Unspecified | 194 |UnrealEngine | 195 |SMPL-H | 196 |
Lens Type | 199 |Fisheye | 200 |Fisheye | 201 |Fisheye | 202 |Rectilinear | 203 |
Mono/Stereo | 206 |Mono | 207 |Mono | 208 |Stereo | 209 |Stereo | 210 |
Body Shape GT | 213 |214 | | 215 | | 216 | | ✓ | 217 |
Joint Location GT | 220 |✓ | 221 |✓ | 222 |✓ | 223 |✓ | 224 |
Joint Rotation GT | 227 |228 | | 229 | | 230 | | ✓ | 231 |
Realism | 234 |Low | 235 |Medium | 236 |High | 237 |High | 238 |
243 | We position the camera on the front of the forehead looking down at the body. 244 | The camera uses a pinhole model approximating the ZED 245 | mini stereo. 246 | We add uniform noise within ±1 cm to the location and ±10° around all axes of 247 | rotation of the camera to simulate misplacement and movement of the HMD on the head. 248 | The resulting images are typically quite challenging for pose estimation, as many parts of the body 249 | are often not seen by the camera. 250 |
251 |261 | The goal of our method is to estimate the probability distribution over joint rotations $\mathbf{R} 262 | = \{\mathbf{R}_i\}^{N}_{i=1}$ conditioned on input image data $\mathbf{X}$, 263 | $p(\mathbf{R}|\mathbf{X})$. 264 | Following Sengupta 265 | et al, we train a neural network to regress Fisher parameters $\mathbf{F} = 266 | \{\mathbf{F}_i\}^{N}_{i=1}$ given input image data $\mathbf{X}$. 267 | From these predicted parameters we can calculate the expected rotation, $\mathbf{\hat{R}}_i$ and the 268 | concentration parameters for each joint $i$, $\kappa_{i,j}$. 269 | The latter describes the uncertainty of the rotation distribution. 270 |
271 | 272 |273 | We train the neural network by minimizing loss $\mathcal{L} = \mathcal{L}_{FNLL} + \mathcal{L}_J$. 274 | $\mathcal{L}_{FNLL}$ is the matrix Fisher negative log-likelihood, promoting accurate local joint 275 | rotations. 276 | $$ 277 | \begin{aligned} 278 | \mathcal{L}_{FNLL}&=\sum_{i=1}^{N}log(c(\mathbf{F}_i))-\text{tr}(\mathbf{F}_i^\top \mathbf{R}_i) 279 | \end{aligned} 280 | $$ 281 | $\mathcal{L}_J$ supervises the 3D joint positions regressed from the parametric body model, SMPL-H, 282 | with shape parameters $\boldsymbol\beta$ with joint regressor $\mathcal{J}$. 283 | $$ 284 | \begin{aligned} 285 | J_{3D}(\mathbf{R},\boldsymbol\beta)=&\mathcal{J}(\textit{SMPL-H}(\mathbf{R}, \boldsymbol\beta))\\ 286 | \mathcal{L}_{J}=&\left \| J_{3D}(\hat{\mathbf{R}},\boldsymbol\beta)- 287 | J_{3D}(\mathbf{R},\boldsymbol\beta) \right \|^2_2 288 | \end{aligned} 289 | $$ 290 | This causes the network to consider the effect of the predicted rotations on the final pose, as the 291 | positions of child joints are influenced by the rotation of their parents in the kinematic tree of 292 | our body model. 293 |
294 |Input | 308 |Method | 309 |PA-MPJPE (mm) | 310 ||||
---|---|---|---|---|---|
Upper Body | 313 |Lower Body | 314 |Hands | 315 |All | 316 |||
Monocular | 321 |xR-EgoPose | 322 |50.18 | 323 |76.76 | 324 |127.34 | 325 |97.48 | 326 |
Ours | 329 |38.48 | 330 |62.35 | 331 |98.94 | 332 |76.05 | 333 ||
Stereo | 336 |UnrealEgo | 337 |48.06 | 338 |77.06 | 339 |117.85 | 340 |91.67 | 341 |
Ours | 344 |34.00 | 345 |54.59 | 346 |87.78 | 347 |67.31 | 348 |
356 | To evaluate the performance on real-world data, we recorded a dataset of 8378 stereo pair images 357 | from 11 different subjects performing actions like squatting, sitting, stretching, crossing 358 | arms, and interacting with small objects. 359 | Overall, our stereo network has the best performance. 360 | We observe that the extra information provided by the right image helps the network to better 361 | predict extremities. 362 | We also note that UnrealEgo and xR-EgoPose perform particularity poorly for lower body joints. 363 | This may be caused by the fact that the legs are not always visible, and that 2D heat-maps 364 | cannot provide uncertainties for joints outside of the image frame. 365 |
366 |Qualitative results of our method compared to recent work for synthetic and real data.
370 |Axis specific concentration for different joints. Concentration is lowest around the primary axis of rotation for a given joint.
377 |380 | Our paper demonstrates that the predicted uncertainty estimates capture extra information and 381 | priors about body pose, and shows empirically that the estimated uncertainties are reliable. 382 | While the former allows us to better explain the prediction of the model, the latter is of 383 | significant importance when it comes to deployment of our method in downstream tasks such as 384 | avatar animation, where uncertainty estimates can be used as a measure of reliance of the 385 | predicted poses. 386 |
387 |Correlation of confidence with error; the higher the confidence the lower the error. Our confidence estimates are therefore reliable for downstream use.
406 |415 | @inproceedings{cuevas2024simpleego, 416 | title={{SimpleEgo}: Predicting probabilistic body pose from egocentric cameras}, 417 | author={Cuevas-Velasquez, Hanz and Hewitt, Charlie and Aliakbarian, Sadegh and Baltru{\v{s}}aitis, Tadas}, 418 | booktitle={2024 International Conference on 3D Vision (3DV)}, 419 | pages={1446--1455}, 420 | year={2024}, 421 | organization={IEEE} 422 | } 423 |424 |