From 9461cf1bee18fae78828ffe87ceebcdf55ca2c8c Mon Sep 17 00:00:00 2001 From: zeppi Date: Tue, 1 Jun 2021 12:19:35 -0400 Subject: [PATCH] cleanup cleanup cleanup --- dist/bundle.es.js | 34 +++++-------------------------- dist/flow-typed/Collections.js | 7 ------- flow-typed/Collections.js | 7 ------- src/redux/actions/claims.js | 9 ++++---- src/redux/actions/collections.js | 20 ++---------------- src/redux/reducers/claims.js | 9 +------- src/redux/reducers/collections.js | 1 - 7 files changed, 13 insertions(+), 74 deletions(-) diff --git a/dist/bundle.es.js b/dist/bundle.es.js index b0cfaa7..bc5a24f 100644 --- a/dist/bundle.es.js +++ b/dist/bundle.es.js @@ -3824,8 +3824,6 @@ function doResolveUris(uris, returnCachedClaims = false, resolveReposts = true) if (collectionIds.length) { dispatch(doFetchItemsInCollections({ collectionIds: collectionIds, pageSize: 5 })); } - // now collection claims are added, get their stuff - // if collections: doResolveCollections(claimIds) return result; }); @@ -4242,7 +4240,6 @@ function doFetchCollectionListMine(page = 1, pageSize = 99999) { collectionIds: items.map(claim => claim.claim_id), page_size: 5 })); - // update or fetch collections? }; const failure = error => { @@ -4287,7 +4284,6 @@ function doClaimSearch(options = { pageSize: options.page_size } }); - // was return true return resolveInfo; }; @@ -4373,6 +4369,10 @@ function doCollectionPublish(options, localId) { params['tags'] = options.tags.map(tag => tag.name); } + if (options.channel_id) { + params['channel_id'] = options.channel_id; + } + return new Promise(resolve => { dispatch({ type: COLLECTION_PUBLISH_STARTED @@ -4657,8 +4657,6 @@ const doFetchItemsInCollections = (resolveItemsOptions, resolveStartedCallback) var _ref = _asyncToGenerator$2(function* (dispatch, getState) { let fetchItemsForCollectionClaim = (() => { var _ref2 = _asyncToGenerator$2(function* (claim, pageSize) { - // take [ {}, {} ], return {} - // only need items [ url... ] and total_items const totalItems = claim.value.claims && claim.value.claims.length; const claimId = claim.claim_id; const itemOrder = claim.value.claims; @@ -4690,21 +4688,8 @@ const doFetchItemsInCollections = (resolveItemsOptions, resolveStartedCallback) }; try { - // sdk had a strange bug that would only return so many, so this had to be batched. - // otherwise large lists of, ~500 channels for a homepage category failed const batchSize = pageSize || FETCH_BATCH_SIZE; const batches = []; - /* - // this was `collection_resolve` which returns claims for collection in order - // however, this fails when a claim is pending. :/ - for (let i = 0; i < Math.ceil(totalItems / batchSize); i++) { - batches[i] = Lbry.collection_resolve({ - claim_id: claimId, - page: i + 1, - page_size: batchSize, - }); - } - */ for (let i = 0; i < Math.ceil(totalItems / batchSize); i++) { batches[i] = lbryProxy.claim_search({ @@ -4826,7 +4811,6 @@ const doFetchItemsInCollections = (resolveItemsOptions, resolveStartedCallback) if (collectionItems) { collectionItems.forEach(function (collectionItem) { - // here's where we would just items.push(collectionItem.permanent_url newItems.push(collectionItem.permanent_url); valueTypes.add(collectionItem.value_type); if (collectionItem.value.stream_type) { @@ -6062,7 +6046,7 @@ function handleClaimAction(state, action) { } else { byId[channel.claim_id] = channel; } - // Also add the permanent_url here until lighthouse returns canonical_url for search results + byUri[channel.permanent_url] = channel.claim_id; byUri[channel.canonical_url] = channel.claim_id; newResolvingUrls.delete(channel.canonical_url); @@ -6076,11 +6060,7 @@ function handleClaimAction(state, action) { byId[collection.claim_id] = collection; } byUri[url] = collection.claim_id; - - // If url isn't a canonical_url, make sure that is added too byUri[collection.canonical_url] = collection.claim_id; - - // Also add the permanent_url here until lighthouse returns canonical_url for search results byUri[collection.permanent_url] = collection.claim_id; newResolvingUrls.delete(collection.canonical_url); newResolvingUrls.delete(collection.permanent_url); @@ -6243,7 +6223,6 @@ reducers[FETCH_COLLECTION_LIST_COMPLETED] = (state, action) => { myCollectionClaimsSet = new Set(state.myCollectionClaims); claims.forEach(claim => { const { canonical_url: canonicalUrl, permanent_url: permanentUrl, claim_id: claimId } = claim; - // maybe add info about items in collection byUri[canonicalUrl] = claimId; byUri[permanentUrl] = claimId; @@ -6532,8 +6511,6 @@ reducers[COLLECTION_PUBLISH_UPDATE_FAILED] = (state, action) => { }); }; -// COLLECTION_PUBLISH_ABANDON_... - reducers[IMPORT_CHANNEL_STARTED] = state => Object.assign({}, state, { pendingChannelImports: true }); reducers[IMPORT_CHANNEL_COMPLETED] = state => Object.assign({}, state, { pendingChannelImports: false }); @@ -7651,7 +7628,6 @@ const collectionsReducer = handleActions({ const newUnpublishedList = Object.assign({}, unpublishedList); const newPendingList = Object.assign({}, pendingList); - const isEdit = editList[claimId]; if (localId) { // new publish newPendingList[claimId] = Object.assign({}, newUnpublishedList[localId] || {}); diff --git a/dist/flow-typed/Collections.js b/dist/flow-typed/Collections.js index c2eba63..927cef7 100644 --- a/dist/flow-typed/Collections.js +++ b/dist/flow-typed/Collections.js @@ -1,10 +1,3 @@ -declare type CollectionUpdateParams = { - remove?: boolean, - claims?: Array, - name?: string, - order?: { from: number, to: number }, -} - declare type Collection = { id: string, items: Array, diff --git a/flow-typed/Collections.js b/flow-typed/Collections.js index c2eba63..927cef7 100644 --- a/flow-typed/Collections.js +++ b/flow-typed/Collections.js @@ -1,10 +1,3 @@ -declare type CollectionUpdateParams = { - remove?: boolean, - claims?: Array, - name?: string, - order?: { from: number, to: number }, -} - declare type Collection = { id: string, items: Array, diff --git a/src/redux/actions/claims.js b/src/redux/actions/claims.js index 88ee81e..e3a47af 100644 --- a/src/redux/actions/claims.js +++ b/src/redux/actions/claims.js @@ -151,8 +151,6 @@ export function doResolveUris( if (collectionIds.length) { dispatch(doFetchItemsInCollections({ collectionIds: collectionIds, pageSize: 5 })); } - // now collection claims are added, get their stuff - // if collections: doResolveCollections(claimIds) return result; } @@ -618,7 +616,6 @@ export function doFetchCollectionListMine(page: number = 1, pageSize: number = 9 page_size: 5, }) ); - // update or fetch collections? }; const failure = error => { @@ -680,7 +677,6 @@ export function doClaimSearch( pageSize: options.page_size, }, }); - // was return true return resolveInfo; }; @@ -762,6 +758,7 @@ export function doCollectionPublish( const params: { name: string, bid: string, + channel_id?: string, blocking?: true, title?: string, thumbnail_url?: string, @@ -786,6 +783,10 @@ export function doCollectionPublish( params['tags'] = options.tags.map(tag => tag.name); } + if (options.channel_id) { + params['channel_id'] = options.channel_id; + } + return new Promise(resolve => { dispatch({ type: ACTIONS.COLLECTION_PUBLISH_STARTED, diff --git a/src/redux/actions/collections.js b/src/redux/actions/collections.js index 39c10c3..a8b7066 100644 --- a/src/redux/actions/collections.js +++ b/src/redux/actions/collections.js @@ -93,7 +93,7 @@ export const doFetchItemsInCollections = ( pageSize?: number, }, resolveStartedCallback?: () => void -) => async (dispatch: Dispatch, getState: GetState) => { +) => async(dispatch: Dispatch, getState: GetState) => { /* 1) make sure all the collection claims are loaded into claims reducer, search/resolve if necessary. 2) get the item claims for each @@ -120,8 +120,6 @@ export const doFetchItemsInCollections = ( const stateAfterClaimSearch = getState(); async function fetchItemsForCollectionClaim(claim: CollectionClaim, pageSize?: number) { - // take [ {}, {} ], return {} - // only need items [ url... ] and total_items const totalItems = claim.value.claims && claim.value.claims.length; const claimId = claim.claim_id; const itemOrder = claim.value.claims; @@ -154,21 +152,8 @@ export const doFetchItemsInCollections = ( }; try { - // sdk had a strange bug that would only return so many, so this had to be batched. - // otherwise large lists of, ~500 channels for a homepage category failed const batchSize = pageSize || FETCH_BATCH_SIZE; const batches: Array> = []; - /* - // this was `collection_resolve` which returns claims for collection in order - // however, this fails when a claim is pending. :/ - for (let i = 0; i < Math.ceil(totalItems / batchSize); i++) { - batches[i] = Lbry.collection_resolve({ - claim_id: claimId, - page: i + 1, - page_size: batchSize, - }); - } - */ for (let i = 0; i < Math.ceil(totalItems / batchSize); i++) { batches[i] = Lbry.claim_search({ @@ -264,7 +249,6 @@ export const doFetchItemsInCollections = ( if (collectionItems) { collectionItems.forEach(collectionItem => { - // here's where we would just items.push(collectionItem.permanent_url newItems.push(collectionItem.permanent_url); valueTypes.add(collectionItem.value_type); if (collectionItem.value.stream_type) { @@ -329,7 +313,7 @@ export const doFetchItemsInCollection = ( return doFetchItemsInCollections(newOptions, cb); }; -export const doCollectionEdit = (collectionId: string, params: CollectionEditParams) => async ( +export const doCollectionEdit = (collectionId: string, params: CollectionEditParams) => async( dispatch: Dispatch, getState: GetState ) => { diff --git a/src/redux/reducers/claims.js b/src/redux/reducers/claims.js index d321f30..e167aac 100644 --- a/src/redux/reducers/claims.js +++ b/src/redux/reducers/claims.js @@ -162,7 +162,7 @@ function handleClaimAction(state: State, action: any): State { } else { byId[channel.claim_id] = channel; } - // Also add the permanent_url here until lighthouse returns canonical_url for search results + byUri[channel.permanent_url] = channel.claim_id; byUri[channel.canonical_url] = channel.claim_id; newResolvingUrls.delete(channel.canonical_url); @@ -176,11 +176,7 @@ function handleClaimAction(state: State, action: any): State { byId[collection.claim_id] = collection; } byUri[url] = collection.claim_id; - - // If url isn't a canonical_url, make sure that is added too byUri[collection.canonical_url] = collection.claim_id; - - // Also add the permanent_url here until lighthouse returns canonical_url for search results byUri[collection.permanent_url] = collection.claim_id; newResolvingUrls.delete(collection.canonical_url); newResolvingUrls.delete(collection.permanent_url); @@ -350,7 +346,6 @@ reducers[ACTIONS.FETCH_COLLECTION_LIST_COMPLETED] = (state: State, action: any): claims.forEach(claim => { const { meta } = claim; const { canonical_url: canonicalUrl, permanent_url: permanentUrl, claim_id: claimId } = claim; - // maybe add info about items in collection byUri[canonicalUrl] = claimId; byUri[permanentUrl] = claimId; @@ -651,8 +646,6 @@ reducers[ACTIONS.COLLECTION_PUBLISH_UPDATE_FAILED] = (state: State, action: any) }); }; -// COLLECTION_PUBLISH_ABANDON_... - reducers[ACTIONS.IMPORT_CHANNEL_STARTED] = (state: State): State => Object.assign({}, state, { pendingChannelImports: true }); diff --git a/src/redux/reducers/collections.js b/src/redux/reducers/collections.js index e228d6d..acd4fa7 100644 --- a/src/redux/reducers/collections.js +++ b/src/redux/reducers/collections.js @@ -95,7 +95,6 @@ const collectionsReducer = handleActions( const newUnpublishedList = Object.assign({}, unpublishedList); const newPendingList = Object.assign({}, pendingList); - const isEdit = editList[claimId]; if (localId) { // new publish newPendingList[claimId] = Object.assign({}, newUnpublishedList[localId] || {});