Permalink
Switch branches/tags
Commits on Nov 16, 2014
Commits on Nov 14, 2014
Commits on Sep 22, 2014
  1. Merge pull request #629 from basho/feature/aae-exchange-bfs-2.0

    jtuple committed Sep 22, 2014
    Reviewed-by: jonmeredith
  2. Add breadth-first AAE exchange

    jtuple committed Aug 28, 2014
    This commit ports the breath-first exchange algorithm from the
    synctree code in riak_ensemble to the hashtree code here in riak_core.
    This level-by-level exchange is necessary to support streaming and/or
    pipelined AAE exchange -- an approach that will soon underpin the new
    AAE-based fullsync replication protocol in Riak.
Commits on Sep 17, 2014
  1. Merge pull request #627 from basho/feature/handoff-sender-sync-tick20…

    borshop committed Sep 17, 2014
    …-rebased
    
    Handoff sender sends sync periodically
    
    Reviewed-by: kellymclaughlin
Commits on Sep 15, 2014
  1. Merge pull request #626 from basho/bugfix/cherrypick/handoff-fold-abo…

    borshop committed Sep 15, 2014
    …rt-early-2.0
    
    Allow handoff sender to abort handoff by throw'ing from fold fun
    
    Reviewed-by: andrewjstone
  2. Allow handoff sender to abort handoff by throw'ing from fold fun

    slfritchie authored and macintux committed Mar 17, 2014
    There's no reason why the handoff sender fold fun has to continue
    folding when there's a TCP error.  Instead, throw the #ho_acc{}
    record with its error={error,Reason} field intact.
    
    If we force the handoff receiver to exit (using exit(HO_pid, kill)),
    then we simply see the following logged on the sender side:
    
        2014-03-17 15:53:30.949 [error] <0.27207.2>@riak_core_handoff_sender:start_fold:266 hinted_handoff transfer of riak_kv_vnode from 'dev1@127.0.0.1' 1415829711164312202009819681693899175291684651008 to 'dev2@127.0.0.1' 1415829711164312202009819681693899175291684651008 failed because of closed
        2014-03-17 15:53:30.950 [error] <0.220.0>@riak_core_handoff_manager:handle_info:289 An outbound handoff of partition riak_kv_vnode 1415829711164312202009819681693899175291684651008 was terminated for reason: {shutdown,{error,closed}}
    
    Also, change the lager:info() message when the handoff receiver
    gets a tcp_error message that *includes* the reason and
    distiguishes itself from other handoff receiver log messages.
    
    Also, one whitespace correction.
Commits on Sep 4, 2014
  1. Set rebar.config deps back to 2.0 branches

    Jared Morrow
    Jared Morrow committed Sep 4, 2014
Commits on Aug 29, 2014
  1. Update dependencies to point to tags for 2.0

    Jared Morrow
    Jared Morrow committed Aug 29, 2014
Commits on Aug 13, 2014
  1. Merge pull request #621 from basho/bug/jrw/missing-vnode-mgmt-timer-m…

    borshop committed Aug 13, 2014
    …apping
    
    add cuttlefish mapping for vnode_management_timer
    
    Reviewed-by: seancribbs
Commits on Aug 6, 2014
Commits on Jul 23, 2014
  1. Update all branch pointers to 2.0 branches

    Jared Morrow
    Jared Morrow committed Jul 23, 2014
Commits on Jul 11, 2014
  1. Merge pull request #612 from basho/feature/tools.mk-0.5.5

    borshop committed Jul 11, 2014
    Update tools.mk to v0.5.5
    
    Reviewed-by: russelldb
Commits on Jul 10, 2014
  1. Merge pull request #609 from basho/bugfix/use-proxy-to-prevent-crash-…

    borshop committed Jul 10, 2014
    …when-joining-sc-nodes-mixed-clusters
    
    Use a proxy process in claimant when joining/removing SC nodes
    
    Reviewed-by: lordnull
Commits on Jul 8, 2014
Commits on Jul 3, 2014
  1. Merge pull request #611 from basho/feature/tools.mk-0.5.4

    borshop committed Jul 3, 2014
    Update tools.mk to v0.5.4
    
    Reviewed-by: seancribbs
  2. Update tools.mk to v0.5.4

    reiddraper committed Jul 3, 2014
    Removes the need to specify line numbers in the dialyzer ignore file.
Commits on Jul 2, 2014
  1. Demonitor on success in proxy_spawn/1

    andrewjstone committed Jul 2, 2014
    When we spawn a proxy process to run a function and if it successfully
    returns we need to demonitor it so that we don't receive errant 'DOWN'
    messages.
Commits on Jul 1, 2014
  1. Use a proxy process when joining/removing SC nodes

    andrewjstone committed Jun 30, 2014
    Centralize the proxy process implementation in riak_core_util and
    utilize that in riak_core_claimant:bootsrap_members/1.
    
    We use a proxy process for riak_ensemble_manager:{join/remove} to handle
    any errors resulting from the riak_ensemble_manager crashing, and
    to prevent late responses from getting sent to the claimant in the case
    that it already got a timeout for the given operation.
    
    Also update dialyzer.ignore-warnings.
Commits on Jun 20, 2014
  1. Merge pull request #606 from basho/feature/silence-core-vnode-eqc

    borshop committed Jun 20, 2014
    Silence output from core_vnode_eqc and log to file instead
    
    Reviewed-by: seancribbs
  2. Merge pull request #605 from basho/bugfix/handle-sc-vnode-overload

    borshop committed Jun 20, 2014
    call Mod:handle_overload_info/2 for unknown msgs in vnode_proxy during o...
    
    Reviewed-by: kellymclaughlin
  3. Stub out fakemod:handle_overload_info/2 for test

    andrewjstone committed Jun 20, 2014
    In riak_core_vnode_proxy:overload_test_/0 we use fakemod as a vnode
    module because we don't expect any callbacks from the proxy, even in
    overload. This prior commit however, ensures that the vnode will get
    called back with any messages that aren't handled directly via the
    proxy. Previously those messages were dropped on the floor.
Commits on Jun 19, 2014
Commits on Jun 13, 2014
  1. Merge pull request #603 from basho/bugfix/reip-update-claimant

    borshop committed Jun 13, 2014
    Bugfix/reip update claimant
    
    Reviewed-by: reiddraper
Commits on Jun 12, 2014
  1. Merge pull request #602 from basho/bugfix/pmap-stall

    borshop committed Jun 12, 2014
    Fix riak_core_util:pmap/2 infinite stall
    
    Reviewed-by: engelsanchez
  2. Convert members and seen in rename_node from list to orddict, fixing …

    jcapricebasho committed Jun 12, 2014
    …broken orddict:find calls after a reip
Commits on Jun 11, 2014
  1. Fix riak_core_util:pmap/2 infinite stall

    jtuple committed Jun 11, 2014
    This commit changes riak_core_util:pmap/2 to use spawn_link rather
    than spawn to create the asynchronous processes. Without this change,
    pmap/2 can stall forever if any of these processes crashes -- eg. by
    using a map function that generates an error.
    
    This commit also fixes the function -spec as well as optimizes the
    final stage of the pmap by using a list comprehension rather than
    lists:unzip (this is faster and generates less garbage).
  2. Update rename_node to also set claimant field.

    jonmeredith committed Jun 11, 2014
    If reip is being used to restore a dead cluster and the previous
    claimant node is renamed first then the ring never converges (see below).
    This change modifies rename_node to also check the claimant field
    and update as necessary.  The ring manager / claimant could also
    be hardened to verify the claimant is a valid node in the cluster,
    but I suspect that will be worked with riak_ensemble in the near future.
    
    Example with previous claimant dev1 renamed to dev4
    
    red:dev john$ dev1/bin/riak-admin ring-status
    ================================== Claimant ===================================
    Claimant:  'dev1@127.0.0.1'
    Status:    down
    Ring Ready: unknown
    
    ============================== Ownership Handoff ==============================
    No pending changes.
    
    ============================== Unreachable Nodes ==============================
    The following nodes are unreachable: ['dev1@127.0.0.1','dev2@127.0.0.1',
                                          'dev3@127.0.0.1']
    
    WARNING: The cluster state will not converge until all nodes
    are up. Once the above nodes come back online, convergence
    will continue. If the outages are long-term or permanent, you
    can either mark the nodes as down (riak-admin down NODE) or
    forcibly remove the nodes from the cluster (riak-admin
    force-remove NODE) to allow the remaining nodes to settle.
    red:dev john$ dev1/bin/riak-admin down dev3@127.0.0.1
    Failed: "dev3@127.0.0.1" is not a member of the cluster.