From 9461cf1bee18fae78828ffe87ceebcdf55ca2c8c Mon Sep 17 00:00:00 2001
From: zeppi <jessopb@gmail.com>
Date: Tue, 1 Jun 2021 12:19:35 -0400
Subject: [PATCH] cleanup

cleanup

cleanup
---
 dist/bundle.es.js                 | 34 +++++--------------------------
 dist/flow-typed/Collections.js    |  7 -------
 flow-typed/Collections.js         |  7 -------
 src/redux/actions/claims.js       |  9 ++++----
 src/redux/actions/collections.js  | 20 ++----------------
 src/redux/reducers/claims.js      |  9 +-------
 src/redux/reducers/collections.js |  1 -
 7 files changed, 13 insertions(+), 74 deletions(-)

diff --git a/dist/bundle.es.js b/dist/bundle.es.js
index b0cfaa7..bc5a24f 100644
--- a/dist/bundle.es.js
+++ b/dist/bundle.es.js
@@ -3824,8 +3824,6 @@ function doResolveUris(uris, returnCachedClaims = false, resolveReposts = true)
         if (collectionIds.length) {
           dispatch(doFetchItemsInCollections({ collectionIds: collectionIds, pageSize: 5 }));
         }
-        // now collection claims are added, get their stuff
-        // if collections: doResolveCollections(claimIds)
 
         return result;
       });
@@ -4242,7 +4240,6 @@ function doFetchCollectionListMine(page = 1, pageSize = 99999) {
         collectionIds: items.map(claim => claim.claim_id),
         page_size: 5
       }));
-      // update or fetch collections?
     };
 
     const failure = error => {
@@ -4287,7 +4284,6 @@ function doClaimSearch(options = {
             pageSize: options.page_size
           }
         });
-        // was return true
         return resolveInfo;
       };
 
@@ -4373,6 +4369,10 @@ function doCollectionPublish(options, localId) {
       params['tags'] = options.tags.map(tag => tag.name);
     }
 
+    if (options.channel_id) {
+      params['channel_id'] = options.channel_id;
+    }
+
     return new Promise(resolve => {
       dispatch({
         type: COLLECTION_PUBLISH_STARTED
@@ -4657,8 +4657,6 @@ const doFetchItemsInCollections = (resolveItemsOptions, resolveStartedCallback)
   var _ref = _asyncToGenerator$2(function* (dispatch, getState) {
     let fetchItemsForCollectionClaim = (() => {
       var _ref2 = _asyncToGenerator$2(function* (claim, pageSize) {
-        // take [ {}, {} ], return {}
-        // only need items [ url... ] and total_items
         const totalItems = claim.value.claims && claim.value.claims.length;
         const claimId = claim.claim_id;
         const itemOrder = claim.value.claims;
@@ -4690,21 +4688,8 @@ const doFetchItemsInCollections = (resolveItemsOptions, resolveStartedCallback)
         };
 
         try {
-          // sdk had a strange bug that would only return so many, so this had to be batched.
-          // otherwise large lists of, ~500 channels for a homepage category failed
           const batchSize = pageSize || FETCH_BATCH_SIZE;
           const batches = [];
-          /*
-            // this was `collection_resolve` which returns claims for collection in order
-            // however, this fails when a claim is pending. :/
-            for (let i = 0; i < Math.ceil(totalItems / batchSize); i++) {
-              batches[i] = Lbry.collection_resolve({
-                claim_id: claimId,
-                page: i + 1,
-                page_size: batchSize,
-              });
-            }
-          */
 
           for (let i = 0; i < Math.ceil(totalItems / batchSize); i++) {
             batches[i] = lbryProxy.claim_search({
@@ -4826,7 +4811,6 @@ const doFetchItemsInCollections = (resolveItemsOptions, resolveStartedCallback)
 
         if (collectionItems) {
           collectionItems.forEach(function (collectionItem) {
-            // here's where we would just items.push(collectionItem.permanent_url
             newItems.push(collectionItem.permanent_url);
             valueTypes.add(collectionItem.value_type);
             if (collectionItem.value.stream_type) {
@@ -6062,7 +6046,7 @@ function handleClaimAction(state, action) {
       } else {
         byId[channel.claim_id] = channel;
       }
-      // Also add the permanent_url here until lighthouse returns canonical_url for search results
+
       byUri[channel.permanent_url] = channel.claim_id;
       byUri[channel.canonical_url] = channel.claim_id;
       newResolvingUrls.delete(channel.canonical_url);
@@ -6076,11 +6060,7 @@ function handleClaimAction(state, action) {
         byId[collection.claim_id] = collection;
       }
       byUri[url] = collection.claim_id;
-
-      // If url isn't a canonical_url, make sure that is added too
       byUri[collection.canonical_url] = collection.claim_id;
-
-      // Also add the permanent_url here until lighthouse returns canonical_url for search results
       byUri[collection.permanent_url] = collection.claim_id;
       newResolvingUrls.delete(collection.canonical_url);
       newResolvingUrls.delete(collection.permanent_url);
@@ -6243,7 +6223,6 @@ reducers[FETCH_COLLECTION_LIST_COMPLETED] = (state, action) => {
     myCollectionClaimsSet = new Set(state.myCollectionClaims);
     claims.forEach(claim => {
       const { canonical_url: canonicalUrl, permanent_url: permanentUrl, claim_id: claimId } = claim;
-      // maybe add info about items in collection
 
       byUri[canonicalUrl] = claimId;
       byUri[permanentUrl] = claimId;
@@ -6532,8 +6511,6 @@ reducers[COLLECTION_PUBLISH_UPDATE_FAILED] = (state, action) => {
   });
 };
 
-// COLLECTION_PUBLISH_ABANDON_...
-
 reducers[IMPORT_CHANNEL_STARTED] = state => Object.assign({}, state, { pendingChannelImports: true });
 
 reducers[IMPORT_CHANNEL_COMPLETED] = state => Object.assign({}, state, { pendingChannelImports: false });
@@ -7651,7 +7628,6 @@ const collectionsReducer = handleActions({
     const newUnpublishedList = Object.assign({}, unpublishedList);
     const newPendingList = Object.assign({}, pendingList);
 
-    const isEdit = editList[claimId];
     if (localId) {
       // new publish
       newPendingList[claimId] = Object.assign({}, newUnpublishedList[localId] || {});
diff --git a/dist/flow-typed/Collections.js b/dist/flow-typed/Collections.js
index c2eba63..927cef7 100644
--- a/dist/flow-typed/Collections.js
+++ b/dist/flow-typed/Collections.js
@@ -1,10 +1,3 @@
-declare type CollectionUpdateParams = {
-  remove?: boolean,
-  claims?: Array<Claim>,
-  name?: string,
-  order?: { from: number, to: number },
-}
-
 declare type Collection = {
   id: string,
   items: Array<?string>,
diff --git a/flow-typed/Collections.js b/flow-typed/Collections.js
index c2eba63..927cef7 100644
--- a/flow-typed/Collections.js
+++ b/flow-typed/Collections.js
@@ -1,10 +1,3 @@
-declare type CollectionUpdateParams = {
-  remove?: boolean,
-  claims?: Array<Claim>,
-  name?: string,
-  order?: { from: number, to: number },
-}
-
 declare type Collection = {
   id: string,
   items: Array<?string>,
diff --git a/src/redux/actions/claims.js b/src/redux/actions/claims.js
index 88ee81e..e3a47af 100644
--- a/src/redux/actions/claims.js
+++ b/src/redux/actions/claims.js
@@ -151,8 +151,6 @@ export function doResolveUris(
         if (collectionIds.length) {
           dispatch(doFetchItemsInCollections({ collectionIds: collectionIds, pageSize: 5 }));
         }
-        // now collection claims are added, get their stuff
-        // if collections: doResolveCollections(claimIds)
 
         return result;
       }
@@ -618,7 +616,6 @@ export function doFetchCollectionListMine(page: number = 1, pageSize: number = 9
           page_size: 5,
         })
       );
-      // update or fetch collections?
     };
 
     const failure = error => {
@@ -680,7 +677,6 @@ export function doClaimSearch(
           pageSize: options.page_size,
         },
       });
-      // was return true
       return resolveInfo;
     };
 
@@ -762,6 +758,7 @@ export function doCollectionPublish(
     const params: {
       name: string,
       bid: string,
+      channel_id?: string,
       blocking?: true,
       title?: string,
       thumbnail_url?: string,
@@ -786,6 +783,10 @@ export function doCollectionPublish(
       params['tags'] = options.tags.map(tag => tag.name);
     }
 
+    if (options.channel_id) {
+      params['channel_id'] = options.channel_id;
+    }
+
     return new Promise(resolve => {
       dispatch({
         type: ACTIONS.COLLECTION_PUBLISH_STARTED,
diff --git a/src/redux/actions/collections.js b/src/redux/actions/collections.js
index 39c10c3..a8b7066 100644
--- a/src/redux/actions/collections.js
+++ b/src/redux/actions/collections.js
@@ -93,7 +93,7 @@ export const doFetchItemsInCollections = (
     pageSize?: number,
   },
   resolveStartedCallback?: () => void
-) => async (dispatch: Dispatch, getState: GetState) => {
+) => async(dispatch: Dispatch, getState: GetState) => {
   /*
   1) make sure all the collection claims are loaded into claims reducer, search/resolve if necessary.
   2) get the item claims for each
@@ -120,8 +120,6 @@ export const doFetchItemsInCollections = (
   const stateAfterClaimSearch = getState();
 
   async function fetchItemsForCollectionClaim(claim: CollectionClaim, pageSize?: number) {
-    // take [ {}, {} ], return {}
-    // only need items [ url... ] and total_items
     const totalItems = claim.value.claims && claim.value.claims.length;
     const claimId = claim.claim_id;
     const itemOrder = claim.value.claims;
@@ -154,21 +152,8 @@ export const doFetchItemsInCollections = (
     };
 
     try {
-      // sdk had a strange bug that would only return so many, so this had to be batched.
-      // otherwise large lists of, ~500 channels for a homepage category failed
       const batchSize = pageSize || FETCH_BATCH_SIZE;
       const batches: Array<Promise<any>> = [];
-      /*
-        // this was `collection_resolve` which returns claims for collection in order
-        // however, this fails when a claim is pending. :/
-        for (let i = 0; i < Math.ceil(totalItems / batchSize); i++) {
-          batches[i] = Lbry.collection_resolve({
-            claim_id: claimId,
-            page: i + 1,
-            page_size: batchSize,
-          });
-        }
-      */
 
       for (let i = 0; i < Math.ceil(totalItems / batchSize); i++) {
         batches[i] = Lbry.claim_search({
@@ -264,7 +249,6 @@ export const doFetchItemsInCollections = (
 
       if (collectionItems) {
         collectionItems.forEach(collectionItem => {
-          // here's where we would just items.push(collectionItem.permanent_url
           newItems.push(collectionItem.permanent_url);
           valueTypes.add(collectionItem.value_type);
           if (collectionItem.value.stream_type) {
@@ -329,7 +313,7 @@ export const doFetchItemsInCollection = (
   return doFetchItemsInCollections(newOptions, cb);
 };
 
-export const doCollectionEdit = (collectionId: string, params: CollectionEditParams) => async (
+export const doCollectionEdit = (collectionId: string, params: CollectionEditParams) => async(
   dispatch: Dispatch,
   getState: GetState
 ) => {
diff --git a/src/redux/reducers/claims.js b/src/redux/reducers/claims.js
index d321f30..e167aac 100644
--- a/src/redux/reducers/claims.js
+++ b/src/redux/reducers/claims.js
@@ -162,7 +162,7 @@ function handleClaimAction(state: State, action: any): State {
       } else {
         byId[channel.claim_id] = channel;
       }
-      // Also add the permanent_url here until lighthouse returns canonical_url for search results
+
       byUri[channel.permanent_url] = channel.claim_id;
       byUri[channel.canonical_url] = channel.claim_id;
       newResolvingUrls.delete(channel.canonical_url);
@@ -176,11 +176,7 @@ function handleClaimAction(state: State, action: any): State {
         byId[collection.claim_id] = collection;
       }
       byUri[url] = collection.claim_id;
-
-      // If url isn't a canonical_url, make sure that is added too
       byUri[collection.canonical_url] = collection.claim_id;
-
-      // Also add the permanent_url here until lighthouse returns canonical_url for search results
       byUri[collection.permanent_url] = collection.claim_id;
       newResolvingUrls.delete(collection.canonical_url);
       newResolvingUrls.delete(collection.permanent_url);
@@ -350,7 +346,6 @@ reducers[ACTIONS.FETCH_COLLECTION_LIST_COMPLETED] = (state: State, action: any):
     claims.forEach(claim => {
       const { meta } = claim;
       const { canonical_url: canonicalUrl, permanent_url: permanentUrl, claim_id: claimId } = claim;
-      // maybe add info about items in collection
 
       byUri[canonicalUrl] = claimId;
       byUri[permanentUrl] = claimId;
@@ -651,8 +646,6 @@ reducers[ACTIONS.COLLECTION_PUBLISH_UPDATE_FAILED] = (state: State, action: any)
   });
 };
 
-// COLLECTION_PUBLISH_ABANDON_...
-
 reducers[ACTIONS.IMPORT_CHANNEL_STARTED] = (state: State): State =>
   Object.assign({}, state, { pendingChannelImports: true });
 
diff --git a/src/redux/reducers/collections.js b/src/redux/reducers/collections.js
index e228d6d..acd4fa7 100644
--- a/src/redux/reducers/collections.js
+++ b/src/redux/reducers/collections.js
@@ -95,7 +95,6 @@ const collectionsReducer = handleActions(
       const newUnpublishedList = Object.assign({}, unpublishedList);
       const newPendingList = Object.assign({}, pendingList);
 
-      const isEdit = editList[claimId];
       if (localId) {
         // new publish
         newPendingList[claimId] = Object.assign({}, newUnpublishedList[localId] || {});