FOCAL++
FOCAL++
Context Encoder
FOCAL++ adopts two forms of intra-task attention on offline context for more robust task inference in context encoder
Range | Mechanism | Description |
---|---|---|
batch-wise | gated attention | adaptively recalibrates the weights on batch-wise samples |
seqeunce-wise | self attention | captures the correlation along the transition seqeunce |

The two parallel attention module are connected by addition to generate the output task embedding
Contrastive Learning
The context encoder is trained through InfoNCE as a query encoder, along with a momentum counterpart as key encoder
where the momentum encoder is updated through moving average and progresses more slowly than origin encoder

Meta Behavior Learning
Similarly, FOCAL++ trains behavior regularized actor and critic, which is decoupled with the training of task encoder

The trained policy can be directly deployed to new tasks with a few transition samples to generate task embeddings
FOCAL++
http://example.com/2024/10/11/FOCAL++/