While using Eventing Functions or Handlers, the following terminologies are used.
Eventing Functions offers a computing paradigm using which developers can handle data changes via the handlers of OnUpdate or OnDelete. Resources are managed at or above the Eventing Function level, and the containing Eventing Function scopes the state of all handlers.
|In the 6.5 release, the handler code is compressed (with the compressed size is limited to 128KB).|
The OnUpdate handler gets called when a document is created or modified. Two major limitations exist. First, if a document is modified several times in a short duration, the calls may be coalesced into a single event due to deduplication. Second, it is not possible to discern between Create and Update operations.
The OnDelete handler gets called when a document is created or modified. Two major limitations exist. First, it is not possible to discern between Expiration and Delete operation. Second, it is not possible to get the value of the document that was just deleted or expired.
The persistent state of a handler is captured in the below external elements, and all states that appears on the execution stack are ephemeral
The metadata bucket (which in the future will eventually be a system collection).
The documents being observed and their extended attributes.
The storage providers bound to the handler.
Couchbase does not store every version of a document permanently. Hence, when a Handler asks for mutation history of a document, it sees a truncated history of the document. However, the final state of a document is always present in all such histories (as the current state is always available in the database).
Similarly, the KV data engine deduplicates multiple mutations made to any individual document rapidly in succession, to ensure highest possible performance. So, when a document mutates rapidly, Handlers may not see all intermediate states, but in all cases, will see the final state of the document.
An abbreviation of convenience of the term Potentially Recursive Mutation. When a Handler manipulates documents in a bucket that serves as the source of mutations to this or any other Handler, a write originated by a Handler will cause a mutation to be seen by itself or another handler. These are called potentially recursive mutations.
Feed Boundary is a time milestone used during an Eventing Function or Handler configuration. Using the Feed Boundary option, you can either invoke a handler on all data mutations available in the cluster (Everything) or choose to invoke a handler during future instances of data mutation, post deployment (From now).
There are two key buckets used by every Eventing Function or Handler: the source bucket and the metadata bucket.
Couchbase Eventing Functions use a bucket to track data mutations. This bucket is termed as the source bucket. The source bucket can be either Couchbase or Ephemeral bucket type. However, memcached bucket types are not supported.
When you are creating a function, you need to specify a source bucket. The handler code watches this bucket via DCP to track data mutations.
|You can use a common source bucket to listen to multiple Eventing Functions running different code.|
When a source bucket is deleted, all deployed functions associated with this source bucket, are undeployed.
After processing the handler code, documents can be stored in a different bucket. For understanding purposes, this bucket can be termed as a destination bucket.
At times, the handler code can trigger data mutations. To avoid cyclic generation of data changes, refer to Bucket Allocation Considerations.
Metadata bucket stores artifacts (or configuration documents) such as, DCP streams, worker allocations, timer information, along with internal checkpoint information.
When you are creating an Eventing Function, ensure that a separate bucket is designated as a metadata bucket. You can use a common metadata bucket across multiple Eventing Functions.
|At any point, refrain from deleting the metadata bucket. Also, ensure that your handler code does not perform a write operation on the metadata bucket.|
If a metadata bucket gets accidentally deleted, then all deployed functions are undeployed and associated indexes get dropped.
The log levels, N1QL Consistency, Workers, and Script Timeout, are additional options available during the Eventing Function definition process.
Log Levels: Use log levels to determine the granularity of log messages.
The available choices are: Info, Error, Debug, Warning, and Trace.
N1QL Consistency: The default consistency level of N1QL statements in the handler.
This controls the consistency level for N1QL statements, but can be set on a per statement basis. The valid values are "None" and "Request".
Workers: Workers the number of worker processes to be started for the handler.
Allows the handler to be scaled up (or vertical scaling). Each worker process supports two fixed threads of execution, however this setting is limited to a maximum of 64 for system optimization purposes. The system automatically generates a warning message if the number of workers exceeds a set threshold based upon cluster resources, however, in this case the handler can still be deployed.
Language compatibility: The language version of the handler for backward compatibility.
If the semantics of a language construct changes in any given release the “language compatibility” setting will ensure an older handler will continue to see the runtime behavior that existed at the time it was authored, until such behavior is deprecated and removed. For example accessing non-existent items from a bucket returns undefined in 6.5 while in 6.0 an exception is thrown.
Script Timeout: Script Timeout provides a timeout option to terminate a non-responsive Function.
The entry points into the handler, e.g. OnUpdate and OnDelete, processing for each mutation must complete from start to finish prior to this specified timeout duration.
An Eventing Function can have no binding, one binding, or several bindings.
You can add bucket bindings via the 'bucket-alias' choice then entering a tuple of: alias-name, bucket-name, and an access level. Where the alias-name that you can use to refer to the bucket from your handler code; the bucket-name is the actual name of the bucket in the cluster; and the access level is either 'read only' or 'read and write'.
|One or more bucket bindings (or bucket-aliases) are mandatory when your handler code performs any bucket related operations.|
Read Only Bindings: A binding with access level of "Read Only" allows reading documents from the bucket, but cannot be used to write (create, update or delete) documents in such a bucket. Attempting to do so will throw a runtime exception.
Read-Write Bindings: A binding with access level of "Read Write" allows both reading and writing (create, update, delete) of documents in the bucket.
These bindings are utilized by the cURL language construct to access external resources. The binding specifies the endpoint, the protocol (http/https), and credentials if necessary. Cookie support can be enabled via the binding if desired when accessing trusted remote nodes. When a URL binding limits access through to be the URL specified or descendants of it. The target of a URL binding should not be a node that belongs to the Couchbase cluster.
You can add URL bindings via the 'URL-alias' choice then entering the following: alias-name, URL, allow cookies setting, and security settings of validate SSL certificate and an auth type of (no auth, basic, bearer, and digest).
The following operations are exposed through the UI, couchbase-cli and REST APIs.
The deploy operation activates an Eventing function or handler. Eventing functions or handlers can be deployed in a cluster.
This operation activates a handler. Source validations are performed, and only valid handlers can be deployed. Deployment transpiles the code and creates the executable artifacts. The source code of an activated handler cannot be edited. Unless a handler is in deployed state, it will not receive or process any events. Deployment creates necessary metadata, spawns worker processes, calculates initial partitions, and initiates checkpointing of processed stream data.
Deployment for DCP observer (or Feed Boundary) has two variations:
Deploy from Start: The Handler will see a deduplicated history of all documents, ending with the current value of each document. Hence, the Handler will see every document in the bucket at least once.
Deploy from Now: The handlers will see mutations from current time. In other words, the Handler will see only documents that mutate after it is deployed.
This operation causes the handler to stop processing events of all types and shuts down the worker processes associated with the handler. It deletes all timers created by the handler being undeployed and their context documents. It releases any runtime resources acquired by the handler. Handlers in undeployed state allow code to be edited. Newly created handlers start in Undeployed state.
This stops all processing associated with a handler including timer callbacks. A handler in paused state can be edited. Handlers in Paused state can be either Resumed or Undeployed.
This continues processing of a handler that was previously Paused. The backlog of mutations that occurred when the handler was paused will now be processed. The backlog of timers that came due when the handler was paused will now fire. Depending on the system capacity and how long the handler was paused, clearing the backlog may take some time before Handler moves on to current mutations and timers.
When a handler is deleted, the source code implementing the handler, all timers, all processing checkpoints and other artifacts in metadata provider is purged. A future handler by the same name has no relation to a prior deleted handler of the same name. Only undeployed handlers can be deleted.
Debug is a special flag on a handler that causes the next event instance received by the handler be trapped and sent to a separate v8 worker with debugging enabled. The debug worker pauses the trapped event processing and opens a TCP port and generates a Chrome Developer Tools URL with a session cookie that can be used to control the debug worker. All other events, except the trapped event instance, continue unencumbered. If the debugged event instance completes execution, another event instance is trapped for debugging, and this continues till debugging is stopped, at which point any trapped instance runs to completion and debug worker becomes passive.
Debugging is convenience feature intended to help during handler development and should not be used in production environments. Debugger does not provide correctness or functionality guarantees.