Blender addon - Scrub timeline from viewport with a shortcut

Overview

Viewport scrub timeline

Move in the timeline directly in viewport and snap to nearest keyframe

Note : This standalone feature will be added in the native add-on Grease Pencil Tools (Starting 2.93)

Download latest


Description

Pop-up a timeline in viewport so you can reduce mouse travel distance (or work full screen) Available in following editors: 3D viewport, Movie Clip and VSE

Viewport scrubbing in time

Viewport timeline : Alt + MMB (middle mouse button) (default)

Use addon preference to customize shortcut, display colors and behavior.

Use designated shortcut in viewport to call the temporary timeline and scrub

Snap to nearest keyframe : While scrubbing, use a modifier key not used as trigger or Right click (Left if you use Right as trigger)


Changelog:

0.7.6

  • Now respect "limit to frame range" option if activated

0.7.5

  • UI: added keyframe display option and reorganise
  • fix: Changed addon pref color to gamma corrected
  • code: refactor for easier merge with GP tools
    • renamed props
    • Separate addon prefs draw and properties

0.7.4

  • fix: bug when HUD is disabled

0.7.3

  • feat: Added native scrub with in timeline editors with addon defined shortcut (Same ops as shift Right mouse):
    • Dopesheet
    • Graph Editor
    • NLA Editor
    • Sequencer
    • ! not in Clip Graph Editor Where the behavior is broken...
  • Added preference option to enable/disable shortcut propagation in timeline editors

0.7.2

  • fix: VSE display now working as expected (HUD on preview window)
  • feat: add bracket style lines to display frame range
  • fix: hide the misleading keyframe icon on start/end frame range

0.7.0

  • feat: Custom Keymapping change:

    • Modal ops to choose shortcut by pressing it
    • Customisable shortcut allowed (both mouse or keyboard)
  • cleanup: Removed all swapping conditions between mouse and key

  • UI: rearrange by category

  • doc: update infos

0.6.2

  • Added support to scrub in VSE and Movie clip editor:
    • Movie clip OK
    • Disable HUD in VSE for now. get wrong screen coordinates in preview window

0.6.1

  • feat: Mouse + modifier shortcut (enabled by default):
    • Automatically change snap key to unused modifiers and click

0.6.0

  • perf: Improved drawing performance (thanks to J.Fran Matheu for his answers on this)

    • Prepare static drawing batches in invoke
    • Use a more appropriate GPU preset for drawing lines
  • UI: Less distractive display

    • Reduced overall lines heights and changed default playhead color to blue.
    • Exposed lines heights in addon prefs (for testing purpose, might stay if interesting to customize)
  • Cleanup: Removed custom timeline placement (top/bottom)

  • fix: Overlapping texts when user ui scale is bigger than default

  • doc: Marked as WIP in bl_infos

0.5.0

  • feat: Display keyframes on timeline

  • feat: Consider objects key (non-GP object)

  • new pref: Consider GP objet key for snapping/display (defaut=True)

  • fix: Allow "empty" scrubbing (when there is no active object)

  • fix: Problem with onion skin auto-hide whne using Esc to go back to init frame

  • doc: Better readme with demo gif

0.4.2

  • fix: display HUD/OSD only in active viewport

  • fix: non blocking error with an uninitialized frame variable

  • disable onion skin during the modal

0.4.1

  • fix: corrected an offset bug in time when viewport used was not leftmost in screen

0.4.0

  • Base text overlay dpi according to user settings

  • Snap mode on left click + continuous press

0.3.0

  • initial commit
Owner
Samuel Bernou
CG artist. Like : Blender3D, drawing stories, Keyboards, coding free tools, video-games, cashew...
Samuel Bernou
Chatbot for the Chatango messaging platform

BroiestBot The baddest bot in the game right now. Uses the ch.py framework for joining Chantango rooms and responding to user messages. Commands If a

Todd Birchard 3 Jan 17, 2022
PyTorch Implementation of "Bridging Pre-trained Language Models and Hand-crafted Features for Unsupervised POS Tagging" (Findings of ACL 2022)

Feature_CRF_AE Feature_CRF_AE provides a implementation of Bridging Pre-trained Language Models and Hand-crafted Features for Unsupervised POS Tagging

Jacob Zhou 6 Apr 29, 2022
Unsupervised text tokenizer for Neural Network-based text generation.

SentencePiece SentencePiece is an unsupervised text tokenizer and detokenizer mainly for Neural Network-based text generation systems where the vocabu

Google 6.4k Jan 01, 2023
NeuTex: Neural Texture Mapping for Volumetric Neural Rendering

NeuTex: Neural Texture Mapping for Volumetric Neural Rendering Paper: https://arxiv.org/abs/2103.00762 Running Run on the provided DTU scene cd run ba

Fanbo Xiang 68 Jan 06, 2023
CoSENT、STS、SentenceBERT

CoSENT_Pytorch 比Sentence-BERT更有效的句向量方案

102 Dec 07, 2022
ConvBERT-Prod

ConvBERT 目录 0. 仓库结构 1. 简介 2. 数据集和复现精度 3. 准备数据与环境 3.1 准备环境 3.2 准备数据 3.3 准备模型 4. 开始使用 4.1 模型训练 4.2 模型评估 4.3 模型预测 5. 模型推理部署 5.1 基于Inference的推理 5.2 基于Serv

yujun 7 Apr 08, 2022
Implementation of some unbalanced loss like focal_loss, dice_loss, DSC Loss, GHM Loss et.al

Implementation of some unbalanced loss for NLP task like focal_loss, dice_loss, DSC Loss, GHM Loss et.al Summary Here is a loss implementation reposit

121 Jan 01, 2023
A PyTorch implementation of the Transformer model in "Attention is All You Need".

Attention is all you need: A Pytorch Implementation This is a PyTorch implementation of the Transformer model in "Attention is All You Need" (Ashish V

Yu-Hsiang Huang 7.1k Jan 05, 2023
Modeling cumulative cases of Covid-19 in the US during the Covid 19 Delta wave using Bayesian methods.

Introduction The goal of this analysis is to find a model that fits the observed cumulative cases of COVID-19 in the US, starting in Mid-July 2021 and

Alexander Keeney 1 Jan 05, 2022
scikit-learn wrappers for Python fastText.

skift scikit-learn wrappers for Python fastText. from skift import FirstColFtClassifier df = pandas.DataFrame([['woof', 0], ['meow', 1]], colu

Shay Palachy 233 Sep 09, 2022
TFIDF-based QA system for AIO2 competition

AIO2 TF-IDF Baseline This is a very simple question answering system, which is developed as a lightweight baseline for AIO2 competition. In the traini

Masatoshi Suzuki 4 Feb 19, 2022
Conditional Transformer Language Model for Controllable Generation

CTRL - A Conditional Transformer Language Model for Controllable Generation Authors: Nitish Shirish Keskar, Bryan McCann, Lav Varshney, Caiming Xiong,

Salesforce 1.7k Dec 28, 2022
Tokenizer - Module python d'analyse syntaxique et de grammaire, tokenization

Tokenizer Le Tokenizer est un analyseur lexicale, il permet, comme Flex and Yacc par exemple, de tokenizer du code, c'est à dire transformer du code e

Manolo 1 Aug 15, 2022
Baseline code for Korean open domain question answering(ODQA)

Open-Domain Question Answering(ODQA)는 다양한 주제에 대한 문서 집합으로부터 자연어 질의에 대한 답변을 찾아오는 task입니다. 이때 사용자 질의에 답변하기 위해 주어지는 지문이 따로 존재하지 않습니다. 따라서 사전에 구축되어있는 Knowl

VUMBLEB 69 Nov 04, 2022
Machine Learning Course Project, IMDB movie review sentiment analysis by lstm, cnn, and transformer

IMDB Sentiment Analysis This is the final project of Machine Learning Courses in Huazhong University of Science and Technology, School of Artificial I

Daniel 0 Dec 27, 2021
原神抽卡记录数据集-Genshin Impact gacha data

提要 持续收集原神抽卡记录中 可以使用抽卡记录导出工具导出抽卡记录的json,将json文件发送至[email protected],我会在清除个人信息后

117 Dec 27, 2022
A PyTorch implementation of VIOLET

VIOLET: End-to-End Video-Language Transformers with Masked Visual-token Modeling A PyTorch implementation of VIOLET Overview VIOLET is an implementati

Tsu-Jui Fu 119 Dec 30, 2022
BERT has a Mouth, and It Must Speak: BERT as a Markov Random Field Language Model

BERT has a Mouth, and It Must Speak: BERT as a Markov Random Field Language Model

303 Dec 17, 2022
Cải thiện Elasticsearch trong bài toán semantic search sử dụng phương pháp Sentence Embeddings

Cải thiện Elasticsearch trong bài toán semantic search sử dụng phương pháp Sentence Embeddings Trong bài viết này mình sẽ sử dụng pretrain model SimCS

Vo Van Phuc 18 Nov 25, 2022
Train BPE with fastBPE, and load to Huggingface Tokenizer.

BPEer Train BPE with fastBPE, and load to Huggingface Tokenizer. Description The BPETrainer of Huggingface consumes a lot of memory when I am training

Lizhuo 1 Dec 23, 2021