2017-11-11 06:34:12 +04:00
|
|
|
(**************************************************************************)
|
|
|
|
(* *)
|
|
|
|
(* Copyright (c) 2014 - 2017. *)
|
|
|
|
(* Dynamic Ledger Solutions, Inc. <contact@tezos.com> *)
|
|
|
|
(* *)
|
|
|
|
(* All rights reserved. No warranty, explicit or implicit, provided. *)
|
|
|
|
(* *)
|
|
|
|
(**************************************************************************)
|
|
|
|
|
2018-01-26 16:10:20 +04:00
|
|
|
open Block_validator_worker_state
|
|
|
|
open Block_validator_errors
|
|
|
|
let invalid_block block error = Invalid_block { block ; error }
|
2017-11-11 06:34:12 +04:00
|
|
|
|
2018-01-26 16:10:20 +04:00
|
|
|
type limits = {
|
2017-11-13 17:25:02 +04:00
|
|
|
protocol_timeout: float ;
|
2018-01-26 16:10:20 +04:00
|
|
|
worker_limits : Worker_types.limits ;
|
2017-11-11 06:34:12 +04:00
|
|
|
}
|
|
|
|
|
2018-01-26 16:10:20 +04:00
|
|
|
module Name = struct
|
|
|
|
type t = unit
|
|
|
|
let encoding = Data_encoding.empty
|
|
|
|
let base = [ "validator.block" ]
|
|
|
|
let pp _ () = ()
|
|
|
|
end
|
2017-11-11 06:34:12 +04:00
|
|
|
|
2018-01-26 16:10:20 +04:00
|
|
|
module Types = struct
|
|
|
|
include Worker_state
|
|
|
|
type state = {
|
|
|
|
protocol_validator: Protocol_validator.t ;
|
|
|
|
limits : limits ;
|
|
|
|
}
|
|
|
|
type parameters = limits * Distributed_db.t
|
|
|
|
let view _state _parameters = ()
|
|
|
|
end
|
2017-11-11 06:34:12 +04:00
|
|
|
|
2018-01-26 16:10:20 +04:00
|
|
|
module Request = struct
|
|
|
|
include Request
|
|
|
|
type 'a t =
|
|
|
|
| Request_validation : {
|
|
|
|
net_db: Distributed_db.net_db ;
|
|
|
|
notify_new_block: State.Block.t -> unit ;
|
|
|
|
canceler: Lwt_canceler.t option ;
|
2018-01-24 15:48:25 +04:00
|
|
|
peer: P2p_peer.Id.t option ;
|
2018-01-26 16:10:20 +04:00
|
|
|
hash: Block_hash.t ;
|
|
|
|
header: Block_header.t ;
|
|
|
|
operations: Operation.t list list ;
|
|
|
|
} -> State.Block.t tzresult t
|
|
|
|
let view
|
|
|
|
: type a. a t -> view
|
|
|
|
= fun (Request_validation { net_db ; peer ; hash }) ->
|
|
|
|
let net_id = net_db |> Distributed_db.net_state |> State.Net.id in
|
|
|
|
{ net_id ; block = hash ; peer = peer }
|
|
|
|
end
|
2017-11-11 06:34:12 +04:00
|
|
|
|
2018-01-26 16:10:20 +04:00
|
|
|
module Worker = Worker.Make (Name) (Event) (Request) (Types)
|
2017-11-11 06:34:12 +04:00
|
|
|
|
2018-01-26 16:10:20 +04:00
|
|
|
type t = Worker.infinite Worker.queue Worker.t
|
|
|
|
type error += Closed = Worker.Closed
|
2017-11-11 06:34:12 +04:00
|
|
|
|
2018-01-26 16:10:20 +04:00
|
|
|
let debug w =
|
|
|
|
Format.kasprintf (fun msg -> Worker.record_event w (Debug msg))
|
2017-11-11 06:34:12 +04:00
|
|
|
|
|
|
|
let check_header
|
2017-11-19 18:07:59 +04:00
|
|
|
(pred: State.Block.t) hash (header: Block_header.t) =
|
|
|
|
let pred_header = State.Block.header pred in
|
2017-11-11 06:34:12 +04:00
|
|
|
fail_unless
|
|
|
|
(Int32.succ pred_header.shell.level = header.shell.level)
|
|
|
|
(invalid_block hash @@
|
|
|
|
Invalid_level { expected = Int32.succ pred_header.shell.level ;
|
|
|
|
found = header.shell.level }) >>=? fun () ->
|
|
|
|
fail_unless
|
|
|
|
Time.(pred_header.shell.timestamp < header.shell.timestamp)
|
|
|
|
(invalid_block hash Non_increasing_timestamp) >>=? fun () ->
|
|
|
|
fail_unless
|
|
|
|
Fitness.(pred_header.shell.fitness < header.shell.fitness)
|
|
|
|
(invalid_block hash Non_increasing_fitness) >>=? fun () ->
|
|
|
|
fail_unless
|
2017-11-19 18:07:59 +04:00
|
|
|
(header.shell.validation_passes =
|
|
|
|
List.length (State.Block.max_number_of_operations pred))
|
2017-11-11 06:34:12 +04:00
|
|
|
(invalid_block hash
|
|
|
|
(Unexpected_number_of_validation_passes header.shell.validation_passes)
|
|
|
|
) >>=? fun () ->
|
|
|
|
return ()
|
|
|
|
|
|
|
|
let assert_no_duplicate_operations block live_operations operation_hashes =
|
|
|
|
fold_left_s (fold_left_s (fun live_operations oph ->
|
|
|
|
fail_when (Operation_hash.Set.mem oph live_operations)
|
|
|
|
(invalid_block block @@ Replayed_operation oph) >>=? fun () ->
|
|
|
|
return (Operation_hash.Set.add oph live_operations)))
|
|
|
|
live_operations operation_hashes >>=? fun _ ->
|
|
|
|
return ()
|
|
|
|
|
|
|
|
let assert_operation_liveness block live_blocks operations =
|
|
|
|
iter_s (iter_s (fun op ->
|
|
|
|
fail_unless
|
|
|
|
(Block_hash.Set.mem op.Operation.shell.branch live_blocks)
|
|
|
|
(invalid_block block @@
|
|
|
|
Outdated_operation { operation = Operation.hash op ;
|
|
|
|
originating_block = op.shell.branch })))
|
|
|
|
operations
|
|
|
|
|
2017-11-14 04:29:19 +04:00
|
|
|
let check_liveness net_state pred hash operations_hashes operations =
|
|
|
|
begin
|
|
|
|
Chain.data net_state >>= fun chain_data ->
|
|
|
|
if State.Block.equal chain_data.current_head pred then
|
|
|
|
Lwt.return (chain_data.live_blocks, chain_data.live_operations)
|
|
|
|
else
|
|
|
|
Chain_traversal.live_blocks
|
|
|
|
pred (State.Block.max_operations_ttl pred)
|
|
|
|
end >>= fun (live_blocks, live_operations) ->
|
2017-11-11 06:34:12 +04:00
|
|
|
assert_no_duplicate_operations
|
|
|
|
hash live_operations operations_hashes >>=? fun () ->
|
|
|
|
assert_operation_liveness hash live_blocks operations >>=? fun () ->
|
|
|
|
return ()
|
|
|
|
|
|
|
|
let apply_block
|
2017-11-14 04:29:19 +04:00
|
|
|
net_state
|
2017-11-11 06:34:12 +04:00
|
|
|
pred (module Proto : State.Registred_protocol.T)
|
|
|
|
hash (header: Block_header.t)
|
|
|
|
operations =
|
|
|
|
let pred_header = State.Block.header pred
|
|
|
|
and pred_hash = State.Block.hash pred in
|
2017-11-19 18:07:59 +04:00
|
|
|
check_header pred hash header >>=? fun () ->
|
|
|
|
iteri2_p
|
|
|
|
(fun i ops max ->
|
|
|
|
fail_unless
|
|
|
|
(List.length ops <= max)
|
|
|
|
(invalid_block hash @@
|
2017-11-19 18:15:03 +04:00
|
|
|
Too_many_operations
|
|
|
|
{ pass = i + 1 ; found = List.length ops ; max }) >>=? fun () ->
|
|
|
|
let max_size = State.Block.max_operation_data_length pred in
|
|
|
|
iter_p (fun op ->
|
|
|
|
let size = Data_encoding.Binary.length Operation.encoding op in
|
|
|
|
fail_unless
|
|
|
|
(size <= max_size)
|
|
|
|
(invalid_block hash @@
|
|
|
|
Oversized_operation
|
|
|
|
{ operation = Operation.hash op ;
|
|
|
|
size ; max = max_size })) ops >>=? fun () ->
|
|
|
|
return ())
|
2017-11-19 18:07:59 +04:00
|
|
|
operations (State.Block.max_number_of_operations pred) >>=? fun () ->
|
2017-11-11 06:34:12 +04:00
|
|
|
let operation_hashes = List.map (List.map Operation.hash) operations in
|
2017-11-14 04:29:19 +04:00
|
|
|
check_liveness net_state pred hash operation_hashes operations >>=? fun () ->
|
2017-11-11 06:34:12 +04:00
|
|
|
map2_s (map2_s begin fun op_hash raw ->
|
|
|
|
Lwt.return (Proto.parse_operation op_hash raw)
|
|
|
|
|> trace (invalid_block hash (Cannot_parse_operation op_hash))
|
|
|
|
end)
|
|
|
|
operation_hashes
|
|
|
|
operations >>=? fun parsed_operations ->
|
|
|
|
State.Block.context pred >>= fun pred_context ->
|
|
|
|
Context.reset_test_network
|
|
|
|
pred_context pred_hash header.shell.timestamp >>= fun context ->
|
|
|
|
(* TODO wrap 'proto_error' into 'block_error' *)
|
|
|
|
Proto.begin_application
|
|
|
|
~predecessor_context:context
|
|
|
|
~predecessor_timestamp:pred_header.shell.timestamp
|
|
|
|
~predecessor_fitness:pred_header.shell.fitness
|
|
|
|
header >>=? fun state ->
|
|
|
|
fold_left_s (fold_left_s (fun state op ->
|
|
|
|
Proto.apply_operation state op >>=? fun state ->
|
|
|
|
return state))
|
|
|
|
state parsed_operations >>=? fun state ->
|
|
|
|
Proto.finalize_block state >>=? fun new_context ->
|
|
|
|
Context.get_protocol new_context.context >>= fun new_protocol ->
|
|
|
|
let expected_proto_level =
|
|
|
|
if Protocol_hash.equal new_protocol Proto.hash then
|
|
|
|
pred_header.shell.proto_level
|
|
|
|
else
|
|
|
|
(pred_header.shell.proto_level + 1) mod 256 in
|
|
|
|
fail_when (header.shell.proto_level <> expected_proto_level)
|
|
|
|
(invalid_block hash @@ Invalid_proto_level {
|
|
|
|
found = header.shell.proto_level ;
|
|
|
|
expected = expected_proto_level ;
|
|
|
|
}) >>=? fun () ->
|
|
|
|
fail_when
|
|
|
|
Fitness.(new_context.fitness <> header.shell.fitness)
|
|
|
|
(invalid_block hash @@ Invalid_fitness {
|
|
|
|
expected = header.shell.fitness ;
|
|
|
|
found = new_context.fitness ;
|
|
|
|
}) >>=? fun () ->
|
|
|
|
let max_operations_ttl =
|
|
|
|
max 0
|
|
|
|
(min
|
|
|
|
((State.Block.max_operations_ttl pred)+1)
|
|
|
|
new_context.max_operations_ttl) in
|
|
|
|
let new_context =
|
|
|
|
{ new_context with max_operations_ttl } in
|
|
|
|
return new_context
|
|
|
|
|
|
|
|
let check_net_liveness net_db hash (header: Block_header.t) =
|
|
|
|
let net_state = Distributed_db.net_state net_db in
|
|
|
|
match State.Net.expiration net_state with
|
|
|
|
| Some eol when Time.(eol <= header.shell.timestamp) ->
|
|
|
|
fail @@ invalid_block hash @@
|
|
|
|
Expired_network { net_id = State.Net.id net_state ;
|
|
|
|
expiration = eol ;
|
|
|
|
timestamp = header.shell.timestamp }
|
|
|
|
| None | Some _ -> return ()
|
|
|
|
|
|
|
|
let get_proto pred hash =
|
|
|
|
State.Block.context pred >>= fun pred_context ->
|
|
|
|
Context.get_protocol pred_context >>= fun pred_protocol_hash ->
|
|
|
|
match State.Registred_protocol.get pred_protocol_hash with
|
|
|
|
| None ->
|
|
|
|
fail (Unavailable_protocol { block = hash ;
|
|
|
|
protocol = pred_protocol_hash })
|
|
|
|
| Some p -> return p
|
|
|
|
|
2018-01-26 16:10:20 +04:00
|
|
|
let on_request
|
|
|
|
: type r. t -> r Request.t -> r tzresult Lwt.t
|
|
|
|
= fun w
|
|
|
|
(Request.Request_validation
|
|
|
|
{ net_db ; notify_new_block ; canceler ;
|
|
|
|
peer ; hash ; header ; operations }) ->
|
|
|
|
let bv = Worker.state w in
|
|
|
|
let net_state = Distributed_db.net_state net_db in
|
|
|
|
State.Block.read_opt net_state hash >>= function
|
|
|
|
| Some block ->
|
|
|
|
debug w "previously validated block %a (after pipe)"
|
|
|
|
Block_hash.pp_short hash ;
|
|
|
|
Protocol_validator.prefetch_and_compile_protocols
|
|
|
|
bv.protocol_validator
|
|
|
|
?peer ~timeout:bv.limits.protocol_timeout
|
|
|
|
block ;
|
|
|
|
return (Ok block)
|
|
|
|
| None ->
|
|
|
|
State.Block.read_invalid net_state hash >>= function
|
|
|
|
| Some { errors } ->
|
|
|
|
return (Error errors)
|
2017-11-11 06:34:12 +04:00
|
|
|
| None ->
|
2018-01-26 16:10:20 +04:00
|
|
|
begin
|
|
|
|
debug w "validating block %a" Block_hash.pp_short hash ;
|
|
|
|
State.Block.read
|
|
|
|
net_state header.shell.predecessor >>=? fun pred ->
|
|
|
|
get_proto pred hash >>=? fun proto ->
|
|
|
|
(* TODO also protect with [Worker.canceler w]. *)
|
|
|
|
Lwt_utils.protect ?canceler begin fun () ->
|
|
|
|
apply_block
|
|
|
|
(Distributed_db.net_state net_db)
|
|
|
|
pred proto hash header operations
|
|
|
|
end
|
|
|
|
end >>= function
|
|
|
|
| Ok result -> begin
|
|
|
|
Worker.protect w begin fun () ->
|
|
|
|
Distributed_db.commit_block
|
|
|
|
net_db hash header operations result
|
|
|
|
end >>=? function
|
|
|
|
| None ->
|
|
|
|
assert false (* should not happen *)
|
|
|
|
| Some block ->
|
|
|
|
Protocol_validator.prefetch_and_compile_protocols
|
|
|
|
bv.protocol_validator
|
|
|
|
?peer ~timeout:bv.limits.protocol_timeout
|
|
|
|
block ;
|
|
|
|
notify_new_block block ;
|
|
|
|
return (Ok block)
|
|
|
|
end
|
|
|
|
(* TODO catch other temporary error (e.g. system errors)
|
|
|
|
and do not 'commit' them on disk... *)
|
|
|
|
| Error [Lwt_utils.Canceled | Unavailable_protocol _] as err ->
|
|
|
|
return err
|
|
|
|
| Error errors ->
|
|
|
|
Worker.protect w begin fun () ->
|
|
|
|
Distributed_db.commit_invalid_block
|
|
|
|
net_db hash header errors
|
|
|
|
end >>=? fun commited ->
|
|
|
|
assert commited ;
|
|
|
|
return (Error errors)
|
2017-11-11 06:34:12 +04:00
|
|
|
|
2018-01-26 16:10:20 +04:00
|
|
|
let on_launch _ _ (limits, db) =
|
2017-11-11 06:34:12 +04:00
|
|
|
let protocol_validator = Protocol_validator.create db in
|
2018-01-26 16:10:20 +04:00
|
|
|
Lwt.return { Types.protocol_validator ; limits }
|
|
|
|
|
|
|
|
let on_error w r st errs =
|
|
|
|
Worker.record_event w (Validation_failure (r, st, errs)) ;
|
|
|
|
Lwt.return (Error errs)
|
|
|
|
|
|
|
|
let on_completion
|
|
|
|
: type a. t -> a Request.t -> a -> Worker_types.request_status -> unit Lwt.t
|
|
|
|
= fun w (Request.Request_validation _ as r) v st ->
|
|
|
|
match v with
|
|
|
|
| Ok _ ->
|
|
|
|
Worker.record_event w
|
|
|
|
(Event.Validation_success (Request.view r, st)) ;
|
|
|
|
Lwt.return ()
|
|
|
|
| Error errs ->
|
|
|
|
Worker.record_event w
|
|
|
|
(Event.Validation_failure (Request.view r, st, errs)) ;
|
|
|
|
Lwt.return ()
|
2017-11-11 06:34:12 +04:00
|
|
|
|
2018-01-26 16:10:20 +04:00
|
|
|
let table = Worker.create_table Queue
|
2017-11-11 06:34:12 +04:00
|
|
|
|
2018-01-26 16:10:20 +04:00
|
|
|
let create limits db =
|
|
|
|
let module Handlers = struct
|
|
|
|
type self = t
|
|
|
|
let on_launch = on_launch
|
|
|
|
let on_request = on_request
|
|
|
|
let on_close _ = Lwt.return ()
|
|
|
|
let on_error = on_error
|
|
|
|
let on_completion = on_completion
|
|
|
|
let on_no_request _ = return ()
|
|
|
|
end in
|
|
|
|
Worker.launch
|
|
|
|
table
|
|
|
|
limits.worker_limits
|
|
|
|
()
|
|
|
|
(limits, db)
|
|
|
|
(module Handlers)
|
|
|
|
|
|
|
|
let shutdown = Worker.shutdown
|
|
|
|
|
|
|
|
let validate w
|
2017-11-11 06:34:12 +04:00
|
|
|
?canceler ?peer ?(notify_new_block = fun _ -> ())
|
|
|
|
net_db hash (header : Block_header.t) operations =
|
2018-01-26 16:10:20 +04:00
|
|
|
let bv = Worker.state w in
|
2017-11-11 06:34:12 +04:00
|
|
|
let net_state = Distributed_db.net_state net_db in
|
|
|
|
State.Block.read_opt net_state hash >>= function
|
|
|
|
| Some block ->
|
2018-01-26 16:10:20 +04:00
|
|
|
debug w "previously validated block %a (before pipe)"
|
|
|
|
Block_hash.pp_short hash ;
|
2017-11-11 06:34:12 +04:00
|
|
|
Protocol_validator.prefetch_and_compile_protocols
|
2018-01-26 16:10:20 +04:00
|
|
|
bv.protocol_validator
|
|
|
|
?peer ~timeout:bv.limits.protocol_timeout
|
2017-11-13 17:25:02 +04:00
|
|
|
block ;
|
2017-11-11 06:34:12 +04:00
|
|
|
return block
|
|
|
|
| None ->
|
|
|
|
map_p (map_p (fun op ->
|
|
|
|
let op_hash = Operation.hash op in
|
|
|
|
return op_hash))
|
|
|
|
operations >>=? fun hashes ->
|
|
|
|
let computed_hash =
|
|
|
|
Operation_list_list_hash.compute
|
|
|
|
(List.map Operation_list_hash.compute hashes) in
|
|
|
|
fail_when
|
|
|
|
(Operation_list_list_hash.compare
|
|
|
|
computed_hash header.shell.operations_hash <> 0)
|
|
|
|
(Inconsistent_operations_hash {
|
|
|
|
block = hash ;
|
|
|
|
expected = header.shell.operations_hash ;
|
|
|
|
found = computed_hash ;
|
|
|
|
}) >>=? fun () ->
|
|
|
|
check_net_liveness net_db hash header >>=? fun () ->
|
2018-01-26 16:10:20 +04:00
|
|
|
Worker.push_request_and_wait w
|
|
|
|
(Request_validation
|
|
|
|
{ net_db ; notify_new_block ; canceler ;
|
|
|
|
peer ; hash ; header ; operations }) >>=? fun result ->
|
|
|
|
Lwt.return result
|
2017-11-11 06:34:12 +04:00
|
|
|
|
2018-01-26 16:10:20 +04:00
|
|
|
let fetch_and_compile_protocol w =
|
|
|
|
let bv = Worker.state w in
|
2017-11-11 06:34:12 +04:00
|
|
|
Protocol_validator.fetch_and_compile_protocol bv.protocol_validator
|
2018-01-26 16:10:20 +04:00
|
|
|
|
|
|
|
let status = Worker.status
|
|
|
|
|
|
|
|
let running_worker () =
|
|
|
|
match Worker.list table with
|
|
|
|
| (_, single) :: _ -> single
|
|
|
|
| [] -> raise Not_found
|
|
|
|
|
|
|
|
let pending_requests t = Worker.pending_requests t
|
|
|
|
|
|
|
|
let current_request t = Worker.current_request t
|
|
|
|
|
|
|
|
let last_events = Worker.last_events
|