Skip to content

Commit 9f22d83

Browse files
committed
Update files
1 parent 48f885c commit 9f22d83

File tree

1 file changed

+280
-0
lines changed

1 file changed

+280
-0
lines changed

json_data/2024/main/image-and-video-synthesis-and-generation.json

Lines changed: 280 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -4198,5 +4198,285 @@
41984198
"onedrive": null,
41994199
"loom": null,
42004200
"section": "Image and Video Synthesis and Generation"
4201+
},
4202+
{
4203+
"title": "Fixed Point Diffusion Models",
4204+
"base_url": "https://openaccess.thecvf.com/content/CVPR2024",
4205+
"title_page": "/html/Bai_Fixed_Point_Diffusion_Models_CVPR_2024_paper.html",
4206+
"ieee_id": null,
4207+
"github": "lukemelas/fixed-point-diffusion-models",
4208+
"web_page": null,
4209+
"github_page": "https://lukemelas.github.io/fixed-point-diffusion-models/",
4210+
"colab": null,
4211+
"modelscope": null,
4212+
"gitee": null,
4213+
"gitlab": null,
4214+
"zenodo": null,
4215+
"kaggle": null,
4216+
"demo_page": null,
4217+
"paper_thecvf": "/papers/Bai_Fixed_Point_Diffusion_Models_CVPR_2024_paper.pdf",
4218+
"paper_arxiv_id": "2401.08741",
4219+
"paper_pdf": null,
4220+
"paper_hal_science": null,
4221+
"paper_researchgate": null,
4222+
"paper_amazon": null,
4223+
"youtube_id": null,
4224+
"drive_google": null,
4225+
"dropbox": null,
4226+
"onedrive": null,
4227+
"loom": null,
4228+
"section": "Image and Video Synthesis and Generation"
4229+
},
4230+
{
4231+
"title": "Focus on Your Instruction: Fine-grained and Multi-Instruction Image Editing by Attention Modulation",
4232+
"base_url": "https://openaccess.thecvf.com/content/CVPR2024",
4233+
"title_page": "/html/Guo_Focus_on_Your_Instruction_Fine-grained_and_Multi-instruction_Image_Editing_by_CVPR_2024_paper.html",
4234+
"ieee_id": null,
4235+
"github": "guoqincode/Focus-on-Your-Instruction",
4236+
"web_page": null,
4237+
"github_page": null,
4238+
"colab": null,
4239+
"modelscope": null,
4240+
"gitee": null,
4241+
"gitlab": null,
4242+
"zenodo": null,
4243+
"kaggle": null,
4244+
"demo_page": null,
4245+
"paper_thecvf": "/papers/Guo_Focus_on_Your_Instruction_Fine-grained_and_Multi-instruction_Image_Editing_by_CVPR_2024_paper.pdf",
4246+
"paper_arxiv_id": "2312.10113",
4247+
"paper_pdf": null,
4248+
"paper_hal_science": null,
4249+
"paper_researchgate": null,
4250+
"paper_amazon": null,
4251+
"youtube_id": "rPknqOJsxkg",
4252+
"drive_google": null,
4253+
"dropbox": null,
4254+
"onedrive": null,
4255+
"loom": null,
4256+
"section": "Image and Video Synthesis and Generation"
4257+
},
4258+
{
4259+
"title": "FreeControl: Training-Free Spatial Control of any Text-to-Image Diffusion Model with any Condition",
4260+
"base_url": "https://openaccess.thecvf.com/content/CVPR2024",
4261+
"title_page": "/html/Mo_FreeControl_Training-Free_Spatial_Control_of_Any_Text-to-Image_Diffusion_Model_with_CVPR_2024_paper.html",
4262+
"ieee_id": null,
4263+
"github": "genforce/freecontrol",
4264+
"web_page": null,
4265+
"github_page": "https://genforce.github.io/freecontrol/",
4266+
"colab": null,
4267+
"modelscope": null,
4268+
"gitee": null,
4269+
"gitlab": null,
4270+
"zenodo": null,
4271+
"kaggle": null,
4272+
"demo_page": null,
4273+
"paper_thecvf": "/papers/Mo_FreeControl_Training-Free_Spatial_Control_of_Any_Text-to-Image_Diffusion_Model_with_CVPR_2024_paper.pdf",
4274+
"paper_arxiv_id": "2312.07536",
4275+
"paper_pdf": null,
4276+
"paper_hal_science": null,
4277+
"paper_researchgate": null,
4278+
"paper_amazon": null,
4279+
"youtube_id": null,
4280+
"drive_google": null,
4281+
"dropbox": null,
4282+
"onedrive": null,
4283+
"loom": null,
4284+
"section": "Image and Video Synthesis and Generation"
4285+
},
4286+
{
4287+
"title": "FreeCustom: Tuning-Free Customized Image Generation for Multi-Concept Composition",
4288+
"base_url": "https://openaccess.thecvf.com/content/CVPR2024",
4289+
"title_page": "/html/Ding_FreeCustom_Tuning-Free_Customized_Image_Generation_for_Multi-Concept_Composition_CVPR_2024_paper.html",
4290+
"ieee_id": null,
4291+
"github": "aim-uofa/FreeCustom",
4292+
"web_page": null,
4293+
"github_page": "https://aim-uofa.github.io/FreeCustom/",
4294+
"colab": null,
4295+
"modelscope": null,
4296+
"gitee": null,
4297+
"gitlab": null,
4298+
"zenodo": null,
4299+
"kaggle": null,
4300+
"demo_page": null,
4301+
"paper_thecvf": "/papers/Ding_FreeCustom_Tuning-Free_Customized_Image_Generation_for_Multi-Concept_Composition_CVPR_2024_paper.pdf",
4302+
"paper_arxiv_id": "2405.13870",
4303+
"paper_pdf": null,
4304+
"paper_hal_science": null,
4305+
"paper_researchgate": null,
4306+
"paper_amazon": null,
4307+
"youtube_id": null,
4308+
"drive_google": null,
4309+
"dropbox": null,
4310+
"onedrive": null,
4311+
"loom": null,
4312+
"section": "Image and Video Synthesis and Generation"
4313+
},
4314+
{
4315+
"title": "FreeDrag: Feature Dragging for Reliable Point-based Image Editing",
4316+
"base_url": "https://openaccess.thecvf.com/content/CVPR2024",
4317+
"title_page": "/html/Ling_FreeDrag_Feature_Dragging_for_Reliable_Point-based_Image_Editing_CVPR_2024_paper.html",
4318+
"ieee_id": null,
4319+
"github": "LPengYang/FreeDrag",
4320+
"web_page": "https://lin-chen.site/projects/freedrag/",
4321+
"github_page": null,
4322+
"colab": null,
4323+
"modelscope": null,
4324+
"gitee": null,
4325+
"gitlab": null,
4326+
"zenodo": null,
4327+
"kaggle": null,
4328+
"demo_page": null,
4329+
"paper_thecvf": "/papers/Ling_FreeDrag_Feature_Dragging_for_Reliable_Point-based_Image_Editing_CVPR_2024_paper.pdf",
4330+
"paper_arxiv_id": "2307.04684",
4331+
"paper_pdf": null,
4332+
"paper_hal_science": null,
4333+
"paper_researchgate": null,
4334+
"paper_amazon": null,
4335+
"youtube_id": "dKPqMQG1CeE",
4336+
"drive_google": null,
4337+
"dropbox": null,
4338+
"onedrive": null,
4339+
"loom": null,
4340+
"section": "Image and Video Synthesis and Generation"
4341+
},
4342+
{
4343+
"title": "FRESCO: Spatial-Temporal Correspondence for Zero-Shot Video Translation",
4344+
"base_url": "https://openaccess.thecvf.com/content/CVPR2024",
4345+
"title_page": "/html/Yang_FRESCO_Spatial-Temporal_Correspondence_for_Zero-Shot_Video_Translation_CVPR_2024_paper.html",
4346+
"ieee_id": null,
4347+
"github": "williamyang1991/FRESCO",
4348+
"web_page": "https://www.mmlab-ntu.com/project/fresco/",
4349+
"github_page": null,
4350+
"colab": null,
4351+
"modelscope": null,
4352+
"gitee": null,
4353+
"gitlab": null,
4354+
"zenodo": null,
4355+
"kaggle": null,
4356+
"demo_page": "https://huggingface.co/spaces/PKUWilliamYang/FRESCO",
4357+
"paper_thecvf": "/papers/Yang_FRESCO_Spatial-Temporal_Correspondence_for_Zero-Shot_Video_Translation_CVPR_2024_paper.pdf",
4358+
"paper_arxiv_id": "2403.12962",
4359+
"paper_pdf": null,
4360+
"paper_hal_science": null,
4361+
"paper_researchgate": null,
4362+
"paper_amazon": null,
4363+
"youtube_id": "jLnGx5H-wLw",
4364+
"drive_google": null,
4365+
"dropbox": null,
4366+
"onedrive": null,
4367+
"loom": null,
4368+
"section": "Image and Video Synthesis and Generation"
4369+
},
4370+
{
4371+
"title": "FSRT: Facial Scene Representation Transformer for Face Reenactment from Factorized Appearance Head-Pose and Facial Expression Features",
4372+
"base_url": "https://openaccess.thecvf.com/content/CVPR2024",
4373+
"title_page": "/html/Rochow_FSRT_Facial_Scene_Representation_Transformer_for_Face_Reenactment_from_Factorized_CVPR_2024_paper.html",
4374+
"ieee_id": null,
4375+
"github": "andrerochow/fsrt",
4376+
"web_page": null,
4377+
"github_page": "https://andrerochow.github.io/fsrt/",
4378+
"colab": null,
4379+
"modelscope": null,
4380+
"gitee": null,
4381+
"gitlab": null,
4382+
"zenodo": null,
4383+
"kaggle": null,
4384+
"demo_page": null,
4385+
"paper_thecvf": "/papers/Rochow_FSRT_Facial_Scene_Representation_Transformer_for_Face_Reenactment_from_Factorized_CVPR_2024_paper.pdf",
4386+
"paper_arxiv_id": "2404.09736",
4387+
"paper_pdf": null,
4388+
"paper_hal_science": null,
4389+
"paper_researchgate": null,
4390+
"paper_amazon": null,
4391+
"youtube_id": "GIoZ8QoshcM",
4392+
"drive_google": null,
4393+
"dropbox": null,
4394+
"onedrive": null,
4395+
"loom": null,
4396+
"section": "Image and Video Synthesis and Generation"
4397+
},
4398+
{
4399+
"title": "Gaussian Shell Maps for Efficient 3D Human Generation",
4400+
"base_url": "https://openaccess.thecvf.com/content/CVPR2024",
4401+
"title_page": "/html/Abdal_Gaussian_Shell_Maps_for_Efficient_3D_Human_Generation_CVPR_2024_paper.html",
4402+
"ieee_id": null,
4403+
"github": "computational-imaging/GSM",
4404+
"web_page": null,
4405+
"github_page": null,
4406+
"colab": null,
4407+
"modelscope": null,
4408+
"gitee": null,
4409+
"gitlab": null,
4410+
"zenodo": null,
4411+
"kaggle": null,
4412+
"demo_page": null,
4413+
"paper_thecvf": "/papers/Abdal_Gaussian_Shell_Maps_for_Efficient_3D_Human_Generation_CVPR_2024_paper.pdf",
4414+
"paper_arxiv_id": "2311.17857",
4415+
"paper_pdf": null,
4416+
"paper_hal_science": null,
4417+
"paper_researchgate": null,
4418+
"paper_amazon": null,
4419+
"youtube_id": "WSTBftn7N3s",
4420+
"drive_google": null,
4421+
"dropbox": null,
4422+
"onedrive": null,
4423+
"loom": null,
4424+
"section": "Image and Video Synthesis and Generation"
4425+
},
4426+
{
4427+
"title": "GaussianDreamer: Fast Generation from Text to 3D Gaussians by Bridging 2D and 3D Diffusion Models",
4428+
"base_url": "https://openaccess.thecvf.com/content/CVPR2024",
4429+
"title_page": "/html/Yi_GaussianDreamer_Fast_Generation_from_Text_to_3D_Gaussians_by_Bridging_CVPR_2024_paper.html",
4430+
"ieee_id": null,
4431+
"github": "hustvl/GaussianDreamer",
4432+
"web_page": "https://taoranyi.com/gaussiandreamer/",
4433+
"github_page": null,
4434+
"colab": null,
4435+
"modelscope": null,
4436+
"gitee": null,
4437+
"gitlab": null,
4438+
"zenodo": null,
4439+
"kaggle": null,
4440+
"demo_page": null,
4441+
"paper_thecvf": "/papers/Yi_GaussianDreamer_Fast_Generation_from_Text_to_3D_Gaussians_by_Bridging_CVPR_2024_paper.pdf",
4442+
"paper_arxiv_id": "2310.08529",
4443+
"paper_pdf": null,
4444+
"paper_hal_science": null,
4445+
"paper_researchgate": null,
4446+
"paper_amazon": null,
4447+
"youtube_id": null,
4448+
"drive_google": null,
4449+
"dropbox": null,
4450+
"onedrive": null,
4451+
"loom": null,
4452+
"section": "Image and Video Synthesis and Generation"
4453+
},
4454+
{
4455+
"title": "GeneAvatar: Generic Expression-Aware Volumetric Head Avatar Editing from a Single Image",
4456+
"base_url": "https://openaccess.thecvf.com/content/CVPR2024",
4457+
"title_page": "/html/Bao_GeneAvatar_Generic_Expression-Aware_Volumetric_Head_Avatar_Editing_from_a_Single_CVPR_2024_paper.html",
4458+
"ieee_id": null,
4459+
"github": "zju3dv/GeneAvatar",
4460+
"web_page": null,
4461+
"github_page": "https://zju3dv.github.io/geneavatar/",
4462+
"colab": null,
4463+
"modelscope": null,
4464+
"gitee": null,
4465+
"gitlab": null,
4466+
"zenodo": null,
4467+
"kaggle": null,
4468+
"demo_page": null,
4469+
"paper_thecvf": "/papers/Bao_GeneAvatar_Generic_Expression-Aware_Volumetric_Head_Avatar_Editing_from_a_Single_CVPR_2024_paper.pdf",
4470+
"paper_arxiv_id": "2404.02152",
4471+
"paper_pdf": null,
4472+
"paper_hal_science": null,
4473+
"paper_researchgate": null,
4474+
"paper_amazon": null,
4475+
"youtube_id": "4zfbfPivtVU",
4476+
"drive_google": null,
4477+
"dropbox": null,
4478+
"onedrive": null,
4479+
"loom": null,
4480+
"section": "Image and Video Synthesis and Generation"
42014481
}
42024482
]

0 commit comments

Comments
 (0)