Skip to content

Commit

Permalink
chore:3月14日
Browse files Browse the repository at this point in the history
  • Loading branch information
aehyok committed Mar 14, 2024
1 parent 92efb11 commit 5227c6c
Show file tree
Hide file tree
Showing 5 changed files with 222 additions and 1 deletion.
3 changes: 3 additions & 0 deletions README.md
Original file line number Diff line number Diff line change
Expand Up @@ -16,6 +16,9 @@
- canva画图 ttps://www.canva.cn
- 查看openai状态
- https://status.openai.com/
## 3月14日
- ffmpeg学习
- https://www.ruanyifeng.com/blog/2020/01/ffmpeg.html
## 3月13日
- https://www.cognition-labs.com/blog
# 3月11日
Expand Down
9 changes: 9 additions & 0 deletions docs/daily/2024-03.md
Original file line number Diff line number Diff line change
@@ -1,3 +1,12 @@
## 3月14日
- ffmpeg学习
- https://blog.csdn.net/yu540135101/article/details/103025957
- https://www.ruanyifeng.com/blog/2020/01/ffmpeg.html
- k8s课程
- IT老齐 https://space.bilibili.com/359351574/search/video?keyword=k8s
- https://www.youtube.com/watch?v=X48VuDVv0do
- window下的windterm https://www.cnblogs.com/cavan2021/p/17317844.html
- mac下为fig
## 3月10日
- OCR 识别
- https://github.com/sdcb/PaddleSharp
Expand Down
48 changes: 47 additions & 1 deletion docs/javascript/2024-03-07-youtobe.md
Original file line number Diff line number Diff line change
Expand Up @@ -60,4 +60,50 @@ https://subtitlestranslator.com/en/
剪映识别youtobe无字幕文件的视频
字幕文件位置 C:\Users\Administrator\Videos


## 将一系列的图片拼接为视频
```
// -r 帧率 默认值为25张,一秒钟拼接25张图片
// -i 指定图片 也可以通过路径
ffmpeg -framerate 1 -i %d.png -c:v libx264 -r 24 -pix_fmt yuv420p output.mp4
```

## 音视频转码(转换文件类型)
```
ffmpeg -i input.mp4 output.avi
```

## 给视频添加水印
```
ffmpeg -i input.mp4 -i watermark.png -filter_complex "overlay=W-w-10:H-h-10" output.mp4
```

## 视频的裁剪
```
// 从第十秒开始,截取到第30秒
ffmpeg -i input.mp4 -ss 00:00:10 -t 00:00:30 -c copy output.mp4
//只保留前60秒
ffmpeg -i input.mp4 -t 60 -c copy output.mp4
```

## 调整音频和视频参数
```
// 调整音频和视频的比特率、帧率、分辨率等参数
ffmpeg -i input.mp4 -b:v 1M -r 30 -s 1280x720 output.mp4
```

## 提取音频文件
```
// -vn 表示不去掉视频
// -c:a copy表示不改变音频编码,直接拷贝
$ ffmpeg -i input.mp4 -vn -c:a copy output.aac
```

## 加速播放视频
```
// 参考链接https://blog.csdn.net/yinshipin007/article/details/127377482
ffmpeg -i FigureAI.mp4 -filter:v "setpts=0.1*PTS" oktou3.mp4
// 同时修改视频和音频
ffmpeg -i FigureAI.mp4 -filter_complex "[0:v]setpts=0.5*PTS[v];[0:a]atempo=2.0[a]" -map "[v]" -map "[a]" output.mp4
```
26 changes: 26 additions & 0 deletions docs/javascript/2024-03-14-FigureAI.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,26 @@
Figure has demonstrated the first fruit of its collaboration with OpenAI to enhance the capabilities of humanoid robots. In a video released today, the Figure 01 bot is seen conversing in real-time.

The development progress at Figure is nothing short of extraordinary. Entrepreneur Brett Adcock only emerged from stealth last year, after gathering together a bunch of key players from Boston Dynamics, Tesla Google DeepMind and Archer Aviation to "create the world's first commercially viable general purpose humanoid robot."

By October, the Figure 01 was already up on its feet and performing basic autonomous tasks. By the turn of the year, the robot had watch-and-learn capabilities, and was ready to enter the workforce at BMW by mid-January.

We got to see it on the warehouse floor last month, just before Figure announced a successful Series B funding round along with a collaboration agreement with OpenAI "to develop next generation AI models for humanoid robots." Now we get a taste for what that means.

Adcock confirmed in an X post that Figure 01's integrated cameras send data to a large vision-language model trained by OpenAI, while Figure's own neural networks also "take images in at 10 Hz through cameras on the robot." OpenAI is also responsible for the ability to understand spoken words, and all of this influx of information is translated into "fast, low level, dexterous robot actions" by Figure's neural net.

He confirmed that the robot was not teleoperated during the demo, and that the video was filmed at actual speed. All up, a remarkable achievement for a partnership that's less than two weeks old – "our goal is to train a world model to operate humanoid robots at the billion-unit level," said Adcock. At this rate, we won't have to wait long.




Figure已展示了其与OpenAI合作增强人形机器人能力的第一个成果。在今天发布的一段视频中,可以看到Figure 01机器人实时对话。

Figure的发展进程无疑是非同寻常的。企业家布雷特·阿德科克仅在去年才从谨慎状态走出,之前他汇集了来自波士顿动力公司、特斯拉、谷歌DeepMind和Archer Aviation的关键人员,目的是"创造世界上第一款商业化可行的通用人形机器人"。

到10月,Figure 01已能独立站立并执行基本任务。转眼至年底,这个机器人就具备了观察学习能力,并于1月中旬准备在宝马公司投入工作。

上个月我们在仓库里见到了它,之后不久Figure就宣布成功完成B轮融资,并与OpenAI达成合作协议"开发下一代人形机器人AI模型"。现在我们终于一窥这意味着什么。

阿德科克在X网站上确认,Figure 01的集成摄像头将数据发送到由OpenAI训练的大型视觉语言模型,而Figure自己的神经网络也"以10Hz频率通过机器人上的摄像头输入图像"。OpenAI还负责理解口语能力,所有这些信息流被Figure的神经网络转化为"快速、低级、灵活的机器人动作"。

他证实,这个演示视频中的机器人没有遥控操作,并且视频是以实际速度拍摄的。总的来说,这是一个令人难以置信的成就,因为这个合作伙伴关系仅维持了不到两周的时间。阿德科克说:"我们的目标是训练一个世界模型,使人形机器人能够百万单位运行。"按照这个速度,我们不用等太久就能看到了。
Loading

0 comments on commit 5227c6c

Please sign in to comment.