- 신규로 표시
- 북마크
- 구독
- 소거
- RSS 피드 구독
- 강조
- 인쇄
- 부적절한 컨텐트 신고
Hello RealSense people,
I've been waiting for this features since I posted the question during the original PercSDK contest over a year ago. So I'm asking again. Does the RealSense Text To Speech API give you phoneme timing maps yet? I'm referring to the timing information you can get with other TTS packages like NeoSpeech, that let you time something like an animated characters mouth movements. That is, a table you get with each Text To Speech generated waveform that tells you, with accuracy, at what millisecond offset a particular phoneme occurs in the generated audio waveform. You can use that information to choose a particular mouth shape for an animated character to "lip sync" the character to the TTS waveform. You didn't have it last year. I really hope you do now! So, is this feature there?
링크가 복사됨
- 신규로 표시
- 북마크
- 구독
- 소거
- RSS 피드 구독
- 강조
- 인쇄
- 부적절한 컨텐트 신고
I hope you don't mind me bumping this up the forum. I'd really like an answer.
Robert
- 신규로 표시
- 북마크
- 구독
- 소거
- RSS 피드 구독
- 강조
- 인쇄
- 부적절한 컨텐트 신고
What we have currently available is in the released SDK.
- 신규로 표시
- 북마크
- 구독
- 소거
- RSS 피드 구독
- 강조
- 인쇄
- 부적절한 컨텐트 신고
Hi Colleen. Are you saying that the current RealSense SDK does have phoneme timing maps unlike the PercSDK?
- 신규로 표시
- 북마크
- 구독
- 소거
- RSS 피드 구독
- 강조
- 인쇄
- 부적절한 컨텐트 신고
I'm saying that the content in the released Gold R1 2014 SDK is what is currently available for the RealSense camera. Items that are there, are marked experimental or beta if they are not yet at Gold. Other items may be on the roadmap.
