{"payload":{"feedbackUrl":"https://github.com/orgs/community/discussions/53140","repo":{"id":634524529,"defaultBranch":"main","name":"x-transformers","ownerLogin":"liujuncn","currentUserCanPush":false,"isFork":true,"isEmpty":false,"createdAt":"2023-04-30T12:13:26.000Z","ownerAvatar":"https://avatars.githubusercontent.com/u/902005?v=4","public":true,"private":false,"isOrgOwned":false},"refInfo":{"name":"","listCacheKey":"v0:1682856813.842655","currentOid":""},"activityList":{"items":[{"before":"25da06ef763f8241081ab5e5b6dbd994a3841ad7","after":"52bcac25437064757d8c4e5bd9e77b9598b462bb","ref":"refs/heads/main","pushedAt":"2023-05-16T03:07:23.029Z","pushType":"push","commitsCount":9,"pusher":{"login":"liujuncn","name":"Liu Jun","path":"/liujuncn","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/902005?s=80&v=4"},"commit":{"message":"move the responsibility for the final norm on the main branch for pre-norm architectures into the AttentionLayers, make sure it is turned off for resiDual and post-norm configs","shortMessageHtmlLink":"move the responsibility for the final norm on the main branch for pre…"}}],"hasNextPage":false,"hasPreviousPage":false,"activityType":"all","actor":null,"timePeriod":"all","sort":"DESC","perPage":30,"cursor":"djE6ks8AAAADLWluqQA","startCursor":null,"endCursor":null}},"title":"Activity · liujuncn/x-transformers"}