{"payload":{"header_redesign_enabled":false,"results":[{"id":"291099685","archived":false,"color":"#DA5B0B","followers":3,"has_funding_file":false,"hl_name":"mshadloo/Neural-Machine-Translation-with-Attention","hl_trunc_description":"I implement encoder-decoder based seq2seq models with attention using Keras. The encoder can be a Bidirectional LSTM, a simple LSTM, or a…","language":"Jupyter Notebook","mirror":false,"owned_by_organization":false,"public":true,"repo":{"repository":{"id":291099685,"name":"Neural-Machine-Translation-with-Attention","owner_id":33298105,"owner_login":"mshadloo","updated_at":"2021-04-20T22:17:53.409Z","has_issues":true}},"sponsorable":false,"topics":["lstm","gru","neural-machine-translation","attention-mechanism","bidirectional-lstm","encoder-decoder-architecture"],"type":"Public","help_wanted_issues_count":0,"good_first_issue_issues_count":0,"starred_by_current_user":false}],"type":"repositories","page":1,"page_count":1,"elapsed_millis":64,"errors":[],"result_count":1,"facets":[],"protected_org_logins":[],"topics":null,"query_id":"","logged_in":false,"sign_up_path":"/signup?source=code_search_results","sign_in_path":"/login?return_to=https%3A%2F%2Fgithub.com%2Fsearch%3Fq%3Drepo%253Amshadloo%252FNeural-Machine-Translation-with-Attention%2B%2Blanguage%253A%2522Jupyter%2BNotebook%2522","metadata":null,"csrf_tokens":{"/mshadloo/Neural-Machine-Translation-with-Attention/star":{"post":"UgfJCV5FCpxmrqVT1R0gWcwFmCfkSBBzwklmzs1bo5actLEIo8VridSrzpgDGdbC4DaxLnhH-Q4Rc58EJF5pJQ"},"/mshadloo/Neural-Machine-Translation-with-Attention/unstar":{"post":"eU_tiJd5d3BE4-Al1w0sguY1d5r0AgygryHoxlF7ruJeXDRqoBEuvlVXH02XQBsixwbhnhNj8YWXqca6p4qx2Q"},"/sponsors/batch_deferred_sponsor_buttons":{"post":"TvUP9onCVNL7fO4TVx0uKFEoBB7vQ173Pt7Xrba_je6jNq1i2bvcZHtNWdGkBaoWeNzx5aGsOwIPnZJL0FgdfQ"}}},"title":"Repository search results"}