diff --git a/dist/bundle.es.js b/dist/bundle.es.js index a2233e7..36ff66b 100644 --- a/dist/bundle.es.js +++ b/dist/bundle.es.js @@ -141,6 +141,10 @@ const CLAIM_REPOST_STARTED = 'CLAIM_REPOST_STARTED'; const CLAIM_REPOST_COMPLETED = 'CLAIM_REPOST_COMPLETED'; const CLAIM_REPOST_FAILED = 'CLAIM_REPOST_FAILED'; const CLEAR_REPOST_ERROR = 'CLEAR_REPOST_ERROR'; +const CHECK_PUBLISH_NAME_STARTED = 'CHECK_PUBLISH_NAME_STARTED'; +const CHECK_PUBLISH_NAME_COMPLETED = 'CHECK_PUBLISH_NAME_COMPLETED'; +const UPDATE_PENDING_CLAIMS = 'UPDATE_PENDING_CLAIMS'; +const UPDATE_CONFIRMED_CLAIMS = 'UPDATE_CONFIRMED_CLAIMS'; // Comments const COMMENT_LIST_STARTED = 'COMMENT_LIST_STARTED'; @@ -413,6 +417,10 @@ var action_types = /*#__PURE__*/Object.freeze({ CLAIM_REPOST_COMPLETED: CLAIM_REPOST_COMPLETED, CLAIM_REPOST_FAILED: CLAIM_REPOST_FAILED, CLEAR_REPOST_ERROR: CLEAR_REPOST_ERROR, + CHECK_PUBLISH_NAME_STARTED: CHECK_PUBLISH_NAME_STARTED, + CHECK_PUBLISH_NAME_COMPLETED: CHECK_PUBLISH_NAME_COMPLETED, + UPDATE_PENDING_CLAIMS: UPDATE_PENDING_CLAIMS, + UPDATE_CONFIRMED_CLAIMS: UPDATE_CONFIRMED_CLAIMS, COMMENT_LIST_STARTED: COMMENT_LIST_STARTED, COMMENT_LIST_COMPLETED: COMMENT_LIST_COMPLETED, COMMENT_LIST_FAILED: COMMENT_LIST_FAILED, @@ -762,7 +770,7 @@ const DONE = 'done'; const READY$1 = 'ready'; const ERROR = 'error'; -var abandon_txo_states = /*#__PURE__*/Object.freeze({ +var abandon_states = /*#__PURE__*/Object.freeze({ PENDING: PENDING, DONE: DONE, READY: READY$1, @@ -2251,7 +2259,7 @@ const makeSelectClaimIsMine = rawUri => { return false; } - return claims && claims[uri] && claims[uri].claim_id && myClaims.has(claims[uri].claim_id); + return claims && claims[uri] && (claims[uri].is_my_output || claims[uri].claim_id && myClaims.has(claims[uri].claim_id)); }); }; @@ -2328,6 +2336,14 @@ const makeSelectCoverForUri = uri => reselect.createSelector(makeSelectClaimForU const selectIsFetchingClaimListMine = reselect.createSelector(selectState$2, state => state.isFetchingClaimListMine); +const selectMyClaimsPage = reselect.createSelector(selectState$2, state => state.myClaimsPageResults || []); + +const selectMyClaimsPageNumber = reselect.createSelector(selectState$2, state => state.claimListMinePage && state.claimListMinePage.items || [], state => state.txoPage && state.txoPage.page || 1); + +const selectMyClaimsPageItemCount = reselect.createSelector(selectState$2, state => state.myClaimsPageTotalResults || 1); + +const selectFetchingMyClaimsPageError = reselect.createSelector(selectState$2, state => state.fetchingClaimListMinePageError); + const selectMyClaims = reselect.createSelector(selectMyActiveClaims, selectClaimsById, selectAbandoningIds, selectPendingClaims, (myClaimIds, byId, abandoningIds, pendingClaims) => { const claims = []; @@ -3145,6 +3161,11 @@ function doResolveUris(uris, returnCachedClaims = false) { return; } + const options = {}; + + if (urisToResolve.length === 1) { + options.include_is_my_output = true; + } dispatch({ type: RESOLVE_URIS_STARTED, data: { uris: normalizedUris } @@ -3152,7 +3173,7 @@ function doResolveUris(uris, returnCachedClaims = false) { const resolveInfo = {}; - lbryProxy.resolve({ urls: urisToResolve }).then(result => { + return lbryProxy.resolve(_extends$5({ urls: urisToResolve }, options)).then(result => { Object.entries(result).forEach(([uri, uriResolveInfo]) => { const fallbackResolveInfo = { stream: null, @@ -3188,6 +3209,7 @@ function doResolveUris(uris, returnCachedClaims = false) { type: RESOLVE_URIS_COMPLETED, data: { resolveInfo } }); + return result; }); }; } @@ -3203,13 +3225,17 @@ function doFetchClaimListMine(page = 1, pageSize = 99999, resolve = true) { }); // $FlowFixMe - lbryProxy.claim_list({ page, page_size: pageSize, claim_type: ['stream', 'repost'], resolve }).then(result => { - const claims = result.items; - + lbryProxy.claim_list({ + page: page, + page_size: pageSize, + claim_type: ['stream', 'repost'], + resolve + }).then(result => { dispatch({ type: FETCH_CLAIM_LIST_MINE_COMPLETED, data: { - claims + result, + resolve } }); }); @@ -3289,7 +3315,7 @@ function doAbandonTxo(txo, cb) { }; } -function doAbandonClaim(txid, nout) { +function doAbandonClaim(txid, nout, cb) { const outpoint = `${txid}:${nout}`; return (dispatch, getState) => { @@ -3322,6 +3348,7 @@ function doAbandonClaim(txid, nout) { message: isClaim ? 'Error abandoning your claim/support' : 'Error unlocking your tip', isError: true })); + if (cb) cb(ERROR); }; const successCallback = () => { @@ -3329,6 +3356,7 @@ function doAbandonClaim(txid, nout) { type: completedActionType, data }); + if (cb) cb(DONE); let abandonMessage; if (isClaim) { @@ -3379,7 +3407,8 @@ function doFetchClaimsByChannel(uri, page = 1) { channel: uri, valid_channel_signature: true, page: page || 1, - order_by: ['release_time'] + order_by: ['release_time'], + include_is_my_output: true }).then(result => { const { items: claims, total_items: claimsInChannel, page: returnedPage } = result; @@ -3622,6 +3651,30 @@ function doRepost(options) { }; } +function doCheckPublishNameAvailability(name) { + return dispatch => { + dispatch({ + type: CHECK_PUBLISH_NAME_STARTED + }); + + return lbryProxy.claim_list({ name: name }).then(result => { + dispatch({ + type: CHECK_PUBLISH_NAME_COMPLETED + }); + if (result.items.length) { + dispatch({ + type: FETCH_CLAIM_LIST_MINE_COMPLETED, + data: { + result, + resolve: false + } + }); + } + return !(result && result.items && result.items.length); + }); + }; +} + function doClearRepostError() { return { type: CLEAR_REPOST_ERROR @@ -3946,13 +3999,10 @@ function doFileList(page = 1, pageSize = 99999) { }; } -function doFetchFileInfosAndPublishedClaims() { +function doFetchFileInfos() { return (dispatch, getState) => { const state = getState(); - const isFetchingClaimListMine = selectIsFetchingClaimListMine(state); const isFetchingFileInfo = selectIsFetchingFileList(state); - - if (!isFetchingClaimListMine) dispatch(doFetchClaimListMine()); if (!isFetchingFileInfo) dispatch(doFileList()); }; } @@ -4352,37 +4402,34 @@ const doPublish = (success, fail) => (dispatch, getState) => { // Calls claim_list_mine until any pending publishes are confirmed const doCheckPendingPublishes = onConfirmed => (dispatch, getState) => { - const state = getState(); - const pendingById = selectPendingById(state); - - if (!Object.keys(pendingById).length) { - return; - } - let publishCheckInterval; const checkFileList = () => { - lbryProxy.stream_list({ page: 1, page_size: 10 }).then(result => { + const state = getState(); + const pendingById = selectPendingById(state); + lbryProxy.claim_list({ page: 1, page_size: 10 }).then(result => { const claims = result.items; - + const claimsToConfirm = []; claims.forEach(claim => { - // If it's confirmed, check if it was pending previously if (claim.confirmations > 0 && pendingById[claim.claim_id]) { delete pendingById[claim.claim_id]; + claimsToConfirm.push(claim); if (onConfirmed) { onConfirmed(claim); } } }); - - dispatch({ - type: FETCH_CLAIM_LIST_MINE_COMPLETED, - data: { - claims - } - }); - - if (!Object.keys(pendingById).length) { + if (claimsToConfirm.length) { + dispatch({ + type: UPDATE_CONFIRMED_CLAIMS, + data: { + claims: claimsToConfirm + } + }); + } + return Object.keys(pendingById).length; + }).then(len => { + if (!len) { clearInterval(publishCheckInterval); } }); @@ -4903,7 +4950,13 @@ const defaultState = { createChannelError: undefined, pendingChannelImport: false, repostLoading: false, - repostError: undefined + repostError: undefined, + fetchingClaimListMinePageError: undefined, + myClaimsPageResults: [], + myClaimsPageNumber: undefined, + myClaimsPageTotalResults: undefined, + isFetchingClaimListMine: false, + isCheckingNameForPublish: false }; function handleClaimAction(state, action) { @@ -4989,16 +5042,22 @@ reducers[FETCH_CLAIM_LIST_MINE_STARTED] = state => Object.assign({}, state, { }); reducers[FETCH_CLAIM_LIST_MINE_COMPLETED] = (state, action) => { - const { claims } = action.data; + const { result, resolve } = action.data; + const claims = result.items; + const page = result.page; + const totalItems = result.total_items; + const byId = Object.assign({}, state.byId); const byUri = Object.assign({}, state.claimsByUri); const pendingById = Object.assign({}, state.pendingById); let myClaimIds = new Set(state.myClaims); + let urlPage = []; claims.forEach(claim => { const uri = buildURI({ streamName: claim.name, streamClaimId: claim.claim_id }); const { claim_id: claimId } = claim; if (claim.type && claim.type.match(/claim|update/)) { + urlPage.push(uri); if (claim.confirmations < 1) { pendingById[claimId] = claim; delete byId[claimId]; @@ -5008,26 +5067,31 @@ reducers[FETCH_CLAIM_LIST_MINE_COMPLETED] = (state, action) => { byUri[uri] = claimId; } myClaimIds.add(claimId); - if (pendingById[claimId] && claim.confirmations > 0) { + if (!resolve && pendingById[claimId] && claim.confirmations > 0) { delete pendingById[claimId]; } } }); - // Remove old pending publishes - Object.values(pendingById) - // $FlowFixMe - .filter(pendingClaim => byId[pendingClaim.claim_id]).forEach(pendingClaim => { + // Remove old pending publishes if resolve if false (resolve=true means confirmations on updates are not 0) + if (!resolve) { + Object.values(pendingById) // $FlowFixMe - delete pendingById[pendingClaim.claim_id]; - }); + .filter(pendingClaim => byId[pendingClaim.claim_id]).forEach(pendingClaim => { + // $FlowFixMe + delete pendingById[pendingClaim.claim_id]; + }); + } return Object.assign({}, state, { isFetchingClaimListMine: false, myClaims: Array.from(myClaimIds), byId, claimsByUri: byUri, - pendingById + pendingById, + myClaimsPageResults: urlPage, + myClaimsPageNumber: page, + myClaimsPageTotalResults: totalItems }); }; @@ -5155,6 +5219,55 @@ reducers[ABANDON_CLAIM_STARTED] = (state, action) => { }); }; +reducers[UPDATE_PENDING_CLAIMS] = (state, action) => { + const { claims } = action.data; + const byId = Object.assign({}, state.byId); + const byUri = Object.assign({}, state.claimsByUri); + const pendingById = Object.assign({}, state.pendingById); + let myClaimIds = new Set(state.myClaims); + + claims.forEach(claim => { + const uri = buildURI({ streamName: claim.name, streamClaimId: claim.claim_id }); + const { claim_id: claimId } = claim; + if (claim.type && claim.type.match(/claim|update/)) { + pendingById[claimId] = claim; + delete byId[claimId]; + byUri[uri] = claimId; + } + myClaimIds.add(claimId); + }); + return Object.assign({}, state, { + myClaims: Array.from(myClaimIds), + byId, + claimsByUri: byUri, + pendingById + }); +}; + +reducers[UPDATE_CONFIRMED_CLAIMS] = (state, action) => { + const { claims } = action.data; + const byId = Object.assign({}, state.byId); + const byUri = Object.assign({}, state.claimsByUri); + const pendingById = Object.assign({}, state.pendingById); + let myClaimIds = new Set(state.myClaims); + + claims.forEach(claim => { + const uri = buildURI({ streamName: claim.name, streamClaimId: claim.claim_id }); + const { claim_id: claimId } = claim; + if (claim.type && claim.type.match(/claim|update/)) { + delete pendingById[claimId]; + byId[claimId] = claim; + } + myClaimIds.add(claimId); + }); + return Object.assign({}, state, { + myClaims: Array.from(myClaimIds), + byId, + claimsByUri: byUri, + pendingById + }); +}; + reducers[ABANDON_CLAIM_SUCCEEDED] = (state, action) => { const { claimId } = action.data; const byId = Object.assign({}, state.byId); @@ -6563,6 +6676,7 @@ const selectChannelIsBlocked = uri => reselect.createSelector(selectBlockedChann return state.includes(uri); }); +exports.ABANDON_STATES = abandon_states; exports.ACTIONS = action_types; exports.CLAIM_VALUES = claim; exports.DAEMON_SETTINGS = daemon_settings; @@ -6580,7 +6694,6 @@ exports.SORT_OPTIONS = sort_options; exports.SPEECH_URLS = speech_urls; exports.THUMBNAIL_STATUSES = thumbnail_upload_statuses; exports.TRANSACTIONS = transaction_types; -exports.TXO_ABANDON_STATES = abandon_txo_states; exports.TXO_LIST = txo_list; exports.TX_LIST = transaction_list; exports.apiCall = apiCall; @@ -6601,6 +6714,7 @@ exports.doBalanceSubscribe = doBalanceSubscribe; exports.doBlurSearchInput = doBlurSearchInput; exports.doCheckAddressIsMine = doCheckAddressIsMine; exports.doCheckPendingPublishes = doCheckPendingPublishes; +exports.doCheckPublishNameAvailability = doCheckPublishNameAvailability; exports.doClaimSearch = doClaimSearch; exports.doClearPublish = doClearPublish; exports.doClearRepostError = doClearRepostError; @@ -6620,7 +6734,7 @@ exports.doFetchChannelListMine = doFetchChannelListMine; exports.doFetchClaimListMine = doFetchClaimListMine; exports.doFetchClaimsByChannel = doFetchClaimsByChannel; exports.doFetchFileInfo = doFetchFileInfo; -exports.doFetchFileInfosAndPublishedClaims = doFetchFileInfosAndPublishedClaims; +exports.doFetchFileInfos = doFetchFileInfos; exports.doFetchTransactions = doFetchTransactions; exports.doFetchTxoPage = doFetchTxoPage; exports.doFileGet = doFileGet; @@ -6770,6 +6884,7 @@ exports.selectFailedPurchaseUris = selectFailedPurchaseUris; exports.selectFetchingClaimSearch = selectFetchingClaimSearch; exports.selectFetchingClaimSearchByQuery = selectFetchingClaimSearchByQuery; exports.selectFetchingMyChannels = selectFetchingMyChannels; +exports.selectFetchingMyClaimsPageError = selectFetchingMyClaimsPageError; exports.selectFetchingTxosError = selectFetchingTxosError; exports.selectFileInfosByOutpoint = selectFileInfosByOutpoint; exports.selectFileInfosDownloaded = selectFileInfosDownloaded; @@ -6798,6 +6913,9 @@ exports.selectMyClaimForUri = selectMyClaimForUri; exports.selectMyClaimUrisWithoutChannels = selectMyClaimUrisWithoutChannels; exports.selectMyClaims = selectMyClaims; exports.selectMyClaimsOutpoints = selectMyClaimsOutpoints; +exports.selectMyClaimsPage = selectMyClaimsPage; +exports.selectMyClaimsPageItemCount = selectMyClaimsPageItemCount; +exports.selectMyClaimsPageNumber = selectMyClaimsPageNumber; exports.selectMyClaimsRaw = selectMyClaimsRaw; exports.selectMyClaimsWithoutChannels = selectMyClaimsWithoutChannels; exports.selectMyStreamUrlsCount = selectMyStreamUrlsCount; diff --git a/src/constants/abandon_txo_states.js b/src/constants/abandon_states.js similarity index 100% rename from src/constants/abandon_txo_states.js rename to src/constants/abandon_states.js diff --git a/src/constants/action_types.js b/src/constants/action_types.js index ed2cb93..deed387 100644 --- a/src/constants/action_types.js +++ b/src/constants/action_types.js @@ -118,6 +118,10 @@ export const CLAIM_REPOST_STARTED = 'CLAIM_REPOST_STARTED'; export const CLAIM_REPOST_COMPLETED = 'CLAIM_REPOST_COMPLETED'; export const CLAIM_REPOST_FAILED = 'CLAIM_REPOST_FAILED'; export const CLEAR_REPOST_ERROR = 'CLEAR_REPOST_ERROR'; +export const CHECK_PUBLISH_NAME_STARTED = 'CHECK_PUBLISH_NAME_STARTED'; +export const CHECK_PUBLISH_NAME_COMPLETED = 'CHECK_PUBLISH_NAME_COMPLETED'; +export const UPDATE_PENDING_CLAIMS = 'UPDATE_PENDING_CLAIMS'; +export const UPDATE_CONFIRMED_CLAIMS = 'UPDATE_CONFIRMED_CLAIMS'; // Comments export const COMMENT_LIST_STARTED = 'COMMENT_LIST_STARTED'; diff --git a/src/index.js b/src/index.js index e41bdf7..1e1b615 100644 --- a/src/index.js +++ b/src/index.js @@ -7,7 +7,7 @@ import * as SORT_OPTIONS from 'constants/sort_options'; import * as THUMBNAIL_STATUSES from 'constants/thumbnail_upload_statuses'; import * as TRANSACTIONS from 'constants/transaction_types'; import * as TX_LIST from 'constants/transaction_list'; -import * as TXO_ABANDON_STATES from 'constants/abandon_txo_states'; +import * as ABANDON_STATES from 'constants/abandon_states'; import * as TXO_LIST from 'constants/txo_list'; import * as SPEECH_URLS from 'constants/speech_urls'; import * as DAEMON_SETTINGS from 'constants/daemon_settings'; @@ -30,7 +30,7 @@ export { TRANSACTIONS, TX_LIST, TXO_LIST, - TXO_ABANDON_STATES, + ABANDON_STATES, SORT_OPTIONS, PAGES, DEFAULT_KNOWN_TAGS, @@ -74,6 +74,7 @@ export { doImportChannel, doRepost, doClearRepostError, + doCheckPublishNameAvailability, } from 'redux/actions/claims'; export { doDeletePurchasedUri, doPurchaseUri, doFileGet } from 'redux/actions/file'; @@ -81,7 +82,7 @@ export { doDeletePurchasedUri, doPurchaseUri, doFileGet } from 'redux/actions/fi export { doFetchFileInfo, doFileList, - doFetchFileInfosAndPublishedClaims, + doFetchFileInfos, doSetFileListSort, } from 'redux/actions/file_info'; @@ -245,6 +246,10 @@ export { selectRepostError, selectRepostLoading, selectClaimIdsByUri, + selectMyClaimsPage, + selectMyClaimsPageNumber, + selectMyClaimsPageItemCount, + selectFetchingMyClaimsPageError, } from 'redux/selectors/claims'; export { makeSelectCommentsForUri } from 'redux/selectors/comments'; diff --git a/src/redux/actions/claims.js b/src/redux/actions/claims.js index 538b853..99e1d1d 100644 --- a/src/redux/actions/claims.js +++ b/src/redux/actions/claims.js @@ -1,6 +1,6 @@ // @flow import * as ACTIONS from 'constants/action_types'; -import * as TXO_STATES from 'constants/abandon_txo_states'; +import * as ABANDON_STATES from 'constants/abandon_states'; import Lbry from 'lbry'; import { normalizeURI } from 'lbryURI'; import { doToast } from 'redux/actions/notifications'; @@ -35,6 +35,11 @@ export function doResolveUris(uris: Array, returnCachedClaims: boolean = return; } + const options: { include_is_my_output?: boolean } = {}; + + if (urisToResolve.length === 1) { + options.include_is_my_output = true; + } dispatch({ type: ACTIONS.RESOLVE_URIS_STARTED, data: { uris: normalizedUris }, @@ -48,7 +53,7 @@ export function doResolveUris(uris: Array, returnCachedClaims: boolean = }, } = {}; - Lbry.resolve({ urls: urisToResolve }).then((result: ResolveResponse) => { + return Lbry.resolve({ urls: urisToResolve, ...options }).then((result: ResolveResponse) => { Object.entries(result).forEach(([uri, uriResolveInfo]) => { const fallbackResolveInfo = { stream: null, @@ -87,6 +92,7 @@ export function doResolveUris(uris: Array, returnCachedClaims: boolean = type: ACTIONS.RESOLVE_URIS_COMPLETED, data: { resolveInfo }, }); + return result; }); }; } @@ -106,24 +112,26 @@ export function doFetchClaimListMine( }); // $FlowFixMe - Lbry.claim_list({ page, page_size: pageSize, claim_type: ['stream', 'repost'], resolve }).then( - (result: StreamListResponse) => { - const claims = result.items; - - dispatch({ - type: ACTIONS.FETCH_CLAIM_LIST_MINE_COMPLETED, - data: { - claims, - }, - }); - } - ); + Lbry.claim_list({ + page: page, + page_size: pageSize, + claim_type: ['stream', 'repost'], + resolve, + }).then((result: StreamListResponse) => { + dispatch({ + type: ACTIONS.FETCH_CLAIM_LIST_MINE_COMPLETED, + data: { + result, + resolve, + }, + }); + }); }; } export function doAbandonTxo(txo: Txo, cb: string => void) { return (dispatch: Dispatch) => { - if (cb) cb(TXO_STATES.PENDING); + if (cb) cb(ABANDON_STATES.PENDING); const isClaim = txo.type === 'claim'; const isSupport = txo.type === 'support' && txo.is_my_input === true; const isTip = txo.type === 'support' && txo.is_my_input === false; @@ -143,7 +151,7 @@ export function doAbandonTxo(txo: Txo, cb: string => void) { }); const errorCallback = () => { - if (cb) cb(TXO_STATES.ERROR); + if (cb) cb(ABANDON_STATES.ERROR); dispatch( doToast({ message: isClaim ? 'Error abandoning your claim/support' : 'Error unlocking your tip', @@ -166,7 +174,7 @@ export function doAbandonTxo(txo: Txo, cb: string => void) { } else { abandonMessage = __('Successfully unlocked your tip!'); } - if (cb) cb(TXO_STATES.DONE); + if (cb) cb(ABANDON_STATES.DONE); dispatch( doToast({ @@ -206,7 +214,7 @@ export function doAbandonTxo(txo: Txo, cb: string => void) { }; } -export function doAbandonClaim(txid: string, nout: number) { +export function doAbandonClaim(txid: string, nout: number, cb: string => void) { const outpoint = `${txid}:${nout}`; return (dispatch: Dispatch, getState: GetState) => { @@ -247,6 +255,7 @@ export function doAbandonClaim(txid: string, nout: number) { isError: true, }) ); + if (cb) cb(ABANDON_STATES.ERROR); }; const successCallback = () => { @@ -254,6 +263,7 @@ export function doAbandonClaim(txid: string, nout: number) { type: completedActionType, data, }); + if (cb) cb(ABANDON_STATES.DONE); let abandonMessage; if (isClaim) { @@ -307,6 +317,7 @@ export function doFetchClaimsByChannel(uri: string, page: number = 1) { valid_channel_signature: true, page: page || 1, order_by: ['release_time'], + include_is_my_output: true, }).then((result: ClaimSearchResponse) => { const { items: claims, total_items: claimsInChannel, page: returnedPage } = result; @@ -583,6 +594,30 @@ export function doRepost(options: StreamRepostOptions) { }; } +export function doCheckPublishNameAvailability(name: string) { + return (dispatch: Dispatch) => { + dispatch({ + type: ACTIONS.CHECK_PUBLISH_NAME_STARTED, + }); + + return Lbry.claim_list({ name: name }).then(result => { + dispatch({ + type: ACTIONS.CHECK_PUBLISH_NAME_COMPLETED, + }); + if (result.items.length) { + dispatch({ + type: ACTIONS.FETCH_CLAIM_LIST_MINE_COMPLETED, + data: { + result, + resolve: false, + }, + }); + } + return !(result && result.items && result.items.length); + }); + }; +} + export function doClearRepostError() { return { type: ACTIONS.CLEAR_REPOST_ERROR, diff --git a/src/redux/actions/file_info.js b/src/redux/actions/file_info.js index b218053..99b4da0 100644 --- a/src/redux/actions/file_info.js +++ b/src/redux/actions/file_info.js @@ -1,7 +1,6 @@ import * as ACTIONS from 'constants/action_types'; import Lbry from 'lbry'; -import { doFetchClaimListMine } from 'redux/actions/claims'; -import { selectClaimsByUri, selectIsFetchingClaimListMine } from 'redux/selectors/claims'; +import { selectClaimsByUri } from 'redux/selectors/claims'; import { selectIsFetchingFileList, selectUrisLoading } from 'redux/selectors/file_info'; export function doFetchFileInfo(uri) { @@ -58,13 +57,10 @@ export function doFileList(page = 1, pageSize = 99999) { }; } -export function doFetchFileInfosAndPublishedClaims() { +export function doFetchFileInfos() { return (dispatch, getState) => { const state = getState(); - const isFetchingClaimListMine = selectIsFetchingClaimListMine(state); const isFetchingFileInfo = selectIsFetchingFileList(state); - - if (!isFetchingClaimListMine) dispatch(doFetchClaimListMine()); if (!isFetchingFileInfo) dispatch(doFileList()); }; } diff --git a/src/redux/actions/publish.js b/src/redux/actions/publish.js index b03693b..c5a0ec0 100644 --- a/src/redux/actions/publish.js +++ b/src/redux/actions/publish.js @@ -359,40 +359,39 @@ export const doCheckPendingPublishes = (onConfirmed: Function) => ( dispatch: Dispatch, getState: GetState ) => { - const state = getState(); - const pendingById = selectPendingById(state); - - if (!Object.keys(pendingById).length) { - return; - } - let publishCheckInterval; const checkFileList = () => { - Lbry.stream_list({ page: 1, page_size: 10 }).then(result => { - const claims = result.items; - - claims.forEach(claim => { - // If it's confirmed, check if it was pending previously - if (claim.confirmations > 0 && pendingById[claim.claim_id]) { - delete pendingById[claim.claim_id]; - if (onConfirmed) { - onConfirmed(claim); + const state = getState(); + const pendingById = selectPendingById(state); + Lbry.claim_list({ page: 1, page_size: 10 }) + .then(result => { + const claims = result.items; + const claimsToConfirm = []; + claims.forEach(claim => { + if (claim.confirmations > 0 && pendingById[claim.claim_id]) { + delete pendingById[claim.claim_id]; + claimsToConfirm.push(claim); + if (onConfirmed) { + onConfirmed(claim); + } } + }); + if (claimsToConfirm.length) { + dispatch({ + type: ACTIONS.UPDATE_CONFIRMED_CLAIMS, + data: { + claims: claimsToConfirm, + }, + }); + } + return Object.keys(pendingById).length; + }) + .then((len) => { + if (!len) { + clearInterval(publishCheckInterval); } }); - - dispatch({ - type: ACTIONS.FETCH_CLAIM_LIST_MINE_COMPLETED, - data: { - claims, - }, - }); - - if (!Object.keys(pendingById).length) { - clearInterval(publishCheckInterval); - } - }); }; publishCheckInterval = setInterval(() => { diff --git a/src/redux/reducers/claims.js b/src/redux/reducers/claims.js index 28caae4..8dbdd28 100644 --- a/src/redux/reducers/claims.js +++ b/src/redux/reducers/claims.js @@ -40,6 +40,12 @@ type State = { pendingChannelImport: string | boolean, repostLoading: boolean, repostError: ?string, + fetchingClaimListMinePageError: ?string, + myClaimsPageResults: Array, + myClaimsPageNumber: ?number, + myClaimsPageTotalResults: ?number, + isFetchingClaimListMine: boolean, + isCheckingNameForPublish: boolean, }; const reducers = {}; @@ -68,6 +74,12 @@ const defaultState = { pendingChannelImport: false, repostLoading: false, repostError: undefined, + fetchingClaimListMinePageError: undefined, + myClaimsPageResults: [], + myClaimsPageNumber: undefined, + myClaimsPageTotalResults: undefined, + isFetchingClaimListMine: false, + isCheckingNameForPublish: false, }; function handleClaimAction(state: State, action: any): State { @@ -162,16 +174,22 @@ reducers[ACTIONS.FETCH_CLAIM_LIST_MINE_STARTED] = (state: State): State => }); reducers[ACTIONS.FETCH_CLAIM_LIST_MINE_COMPLETED] = (state: State, action: any): State => { - const { claims }: { claims: Array } = action.data; + const { result, resolve }: { result: ClaimListResponse, resolve: boolean } = action.data; + const claims = result.items; + const page = result.page; + const totalItems = result.total_items; + const byId = Object.assign({}, state.byId); const byUri = Object.assign({}, state.claimsByUri); const pendingById: { [string]: Claim } = Object.assign({}, state.pendingById); let myClaimIds = new Set(state.myClaims); + let urlPage = []; claims.forEach((claim: Claim) => { const uri = buildURI({ streamName: claim.name, streamClaimId: claim.claim_id }); const { claim_id: claimId } = claim; if (claim.type && claim.type.match(/claim|update/)) { + urlPage.push(uri); if (claim.confirmations < 1) { pendingById[claimId] = claim; delete byId[claimId]; @@ -181,20 +199,22 @@ reducers[ACTIONS.FETCH_CLAIM_LIST_MINE_COMPLETED] = (state: State, action: any): byUri[uri] = claimId; } myClaimIds.add(claimId); - if (pendingById[claimId] && claim.confirmations > 0) { + if (!resolve && pendingById[claimId] && claim.confirmations > 0) { delete pendingById[claimId]; } } }); - // Remove old pending publishes - Object.values(pendingById) - // $FlowFixMe - .filter(pendingClaim => byId[pendingClaim.claim_id]) - .forEach(pendingClaim => { + // Remove old pending publishes if resolve if false (resolve=true means confirmations on updates are not 0) + if (!resolve) { + Object.values(pendingById) // $FlowFixMe - delete pendingById[pendingClaim.claim_id]; - }); + .filter(pendingClaim => byId[pendingClaim.claim_id]) + .forEach(pendingClaim => { + // $FlowFixMe + delete pendingById[pendingClaim.claim_id]; + }); + } return Object.assign({}, state, { isFetchingClaimListMine: false, @@ -202,6 +222,9 @@ reducers[ACTIONS.FETCH_CLAIM_LIST_MINE_COMPLETED] = (state: State, action: any): byId, claimsByUri: byUri, pendingById, + myClaimsPageResults: urlPage, + myClaimsPageNumber: page, + myClaimsPageTotalResults: totalItems, }); }; @@ -338,6 +361,55 @@ reducers[ACTIONS.ABANDON_CLAIM_STARTED] = (state: State, action: any): State => }); }; +reducers[ACTIONS.UPDATE_PENDING_CLAIMS] = (state: State, action: any): State => { + const { claims }: { claims: Array } = action.data; + const byId = Object.assign({}, state.byId); + const byUri = Object.assign({}, state.claimsByUri); + const pendingById: { [string]: Claim } = Object.assign({}, state.pendingById); + let myClaimIds = new Set(state.myClaims); + + claims.forEach((claim: Claim) => { + const uri = buildURI({ streamName: claim.name, streamClaimId: claim.claim_id }); + const { claim_id: claimId } = claim; + if (claim.type && claim.type.match(/claim|update/)) { + pendingById[claimId] = claim; + delete byId[claimId]; + byUri[uri] = claimId; + } + myClaimIds.add(claimId); + }); + return Object.assign({}, state, { + myClaims: Array.from(myClaimIds), + byId, + claimsByUri: byUri, + pendingById, + }); +}; + +reducers[ACTIONS.UPDATE_CONFIRMED_CLAIMS] = (state: State, action: any): State => { + const { claims }: { claims: Array } = action.data; + const byId = Object.assign({}, state.byId); + const byUri = Object.assign({}, state.claimsByUri); + const pendingById: { [string]: Claim } = Object.assign({}, state.pendingById); + let myClaimIds = new Set(state.myClaims); + + claims.forEach((claim: Claim) => { + const uri = buildURI({ streamName: claim.name, streamClaimId: claim.claim_id }); + const { claim_id: claimId } = claim; + if (claim.type && claim.type.match(/claim|update/)) { + delete pendingById[claimId]; + byId[claimId] = claim; + } + myClaimIds.add(claimId); + }); + return Object.assign({}, state, { + myClaims: Array.from(myClaimIds), + byId, + claimsByUri: byUri, + pendingById, + }); +}; + reducers[ACTIONS.ABANDON_CLAIM_SUCCEEDED] = (state: State, action: any): State => { const { claimId }: { claimId: string } = action.data; const byId = Object.assign({}, state.byId); diff --git a/src/redux/selectors/claims.js b/src/redux/selectors/claims.js index ec860a2..a04987c 100644 --- a/src/redux/selectors/claims.js +++ b/src/redux/selectors/claims.js @@ -212,7 +212,11 @@ export const makeSelectClaimIsMine = (rawUri: string) => { return false; } - return claims && claims[uri] && claims[uri].claim_id && myClaims.has(claims[uri].claim_id); + return ( + claims && + claims[uri] && + (claims[uri].is_my_output || (claims[uri].claim_id && myClaims.has(claims[uri].claim_id))) + ); } ); }; @@ -310,8 +314,8 @@ export const makeSelectDateForUri = (uri: string) => (claim.value.release_time ? claim.value.release_time * 1000 : claim.meta && claim.meta.creation_timestamp - ? claim.meta.creation_timestamp * 1000 - : null); + ? claim.meta.creation_timestamp * 1000 + : null); if (!timestamp) { return undefined; } @@ -360,6 +364,28 @@ export const selectIsFetchingClaimListMine = createSelector( state => state.isFetchingClaimListMine ); +export const selectMyClaimsPage = createSelector( + selectState, + state => state.myClaimsPageResults || [] +); + +export const selectMyClaimsPageNumber = createSelector( + selectState, + state => (state.claimListMinePage && state.claimListMinePage.items) || [], + + state => (state.txoPage && state.txoPage.page) || 1 +); + +export const selectMyClaimsPageItemCount = createSelector( + selectState, + state => state.myClaimsPageTotalResults || 1 +); + +export const selectFetchingMyClaimsPageError = createSelector( + selectState, + state => state.fetchingClaimListMinePageError +); + export const selectMyClaims = createSelector( selectMyActiveClaims, selectClaimsById, diff --git a/src/redux/selectors/publish.js b/src/redux/selectors/publish.js index 82adce2..10ed95b 100644 --- a/src/redux/selectors/publish.js +++ b/src/redux/selectors/publish.js @@ -77,10 +77,10 @@ export const selectMyClaimForUri = createSelector( return isStillEditing ? claimsById[editClaimId] : myClaims.find(claim => - !contentName - ? claim.name === claimName - : claim.name === contentName || claim.name === claimName - ); + !contentName + ? claim.name === claimName + : claim.name === contentName || claim.name === claimName + ); } );