/* * Copyright (c) 2009-2014, 2018 Nicira, Inc. * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at: * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ #ifndef DPIF_PROVIDER_H #define DPIF_PROVIDER_H 1 /* Provider interface to dpifs, which provide an interface to an Open vSwitch * datapath. A datapath is a collection of physical or virtual ports that are * exposed over OpenFlow as a single switch. Datapaths and the collections of * ports that they contain may be fixed or dynamic. */ #include "openflow/openflow.h" #include "dpif.h" #include "util.h" #ifdef __cplusplus extern "C" { #endif /* Open vSwitch datapath interface. * * This structure should be treated as opaque by dpif implementations. */ struct dpif { const struct dpif_class *dpif_class; char *base_name; char *full_name; uint8_t netflow_engine_type; uint8_t netflow_engine_id; long long int current_ms; }; struct dpif_ipf_status; struct ipf_dump_ctx; void dpif_init(struct dpif *, const struct dpif_class *, const char *name, uint8_t netflow_engine_type, uint8_t netflow_engine_id); void dpif_uninit(struct dpif *dpif, bool close); static inline void dpif_assert_class(const struct dpif *dpif, const struct dpif_class *dpif_class) { ovs_assert(dpif->dpif_class == dpif_class); } struct dpif_flow_dump { struct dpif *dpif; bool terse; /* If true, key/mask/actions may be omitted. */ }; static inline void dpif_flow_dump_init(struct dpif_flow_dump *dump, const struct dpif *dpif) { dump->dpif = CONST_CAST(struct dpif *, dpif); } struct dpif_flow_dump_thread { struct dpif *dpif; }; static inline void dpif_flow_dump_thread_init(struct dpif_flow_dump_thread *thread, struct dpif_flow_dump *dump) { thread->dpif = dump->dpif; } struct ct_dpif_dump_state; struct ct_dpif_entry; struct ct_dpif_tuple; struct ct_dpif_timeout_policy; enum ct_features; /* 'dpif_ipf_proto_status' and 'dpif_ipf_status' are presently in * sync with 'ipf_proto_status' and 'ipf_status', but more * generally represent a superset of present and future support. */ struct dpif_ipf_proto_status { uint64_t nfrag_accepted; uint64_t nfrag_completed_sent; uint64_t nfrag_expired_sent; uint64_t nfrag_too_small; uint64_t nfrag_overlap; uint64_t nfrag_purged; unsigned int min_frag_size; bool enabled; }; struct dpif_ipf_status { struct dpif_ipf_proto_status v4; struct dpif_ipf_proto_status v6; unsigned int nfrag; unsigned int nfrag_max; }; /* Datapath interface class structure, to be defined by each implementation of * a datapath interface. * * These functions return 0 if successful or a positive errno value on failure, * except where otherwise noted. * * These functions are expected to execute synchronously, that is, to block as * necessary to obtain a result. Thus, they may not return EAGAIN or * EWOULDBLOCK or EINPROGRESS. We may relax this requirement in the future if * and when we encounter performance problems. */ struct dpif_class { /* Type of dpif in this class, e.g. "system", "netdev", etc. * * One of the providers should supply a "system" type, since this is * the type assumed if no type is specified when opening a dpif. */ const char *type; /* If 'true', datapath ports should be destroyed on ofproto destruction. * * This is used by the vswitch at exit, so that it can clean any * datapaths that can not exist without it (e.g. netdev datapath). */ bool cleanup_required; /* If 'true' the specific dpif implementation synchronizes the various * datapath implementation layers, i.e., the dpif's layer in combination * with the underlying netdev offload layers. For example, dpif-netlink * does not sync its kernel flows with the tc ones, i.e., only one gets * installed. On the other hand, dpif-netdev installs both flows, * internally keeps track of both, and represents them as one. */ bool synced_dp_layers; /* Called when the dpif provider is registered, typically at program * startup. Returning an error from this function will prevent any * datapath with this class from being created. * * This function may be set to null if a datapath class needs no * initialization at registration time. */ int (*init)(void); /* Enumerates the names of all known created datapaths (of class * 'dpif_class'), if possible, into 'all_dps'. The caller has already * initialized 'all_dps' and other dpif classes might already have added * names to it. * * This is used by the vswitch at startup, so that it can delete any * datapaths that are not configured. * * Some kinds of datapaths might not be practically enumerable, in which * case this function may be a null pointer. */ int (*enumerate)(struct sset *all_dps, const struct dpif_class *dpif_class); /* Returns the type to pass to netdev_open() when a dpif of class * 'dpif_class' has a port of type 'type', for a few special cases * when a netdev type differs from a port type. For example, when * using the userspace datapath, a port of type "internal" needs to * be opened as "tap". * * Returns either 'type' itself or a string literal, which must not * be freed. */ const char *(*port_open_type)(const struct dpif_class *dpif_class, const char *type); /* Attempts to open an existing dpif called 'name', if 'create' is false, * or to open an existing dpif or create a new one, if 'create' is true. * * 'dpif_class' is the class of dpif to open. * * If successful, stores a pointer to the new dpif in '*dpifp', which must * have class 'dpif_class'. On failure there are no requirements on what * is stored in '*dpifp'. */ int (*open)(const struct dpif_class *dpif_class, const char *name, bool create, struct dpif **dpifp); /* Closes 'dpif' and frees associated memory. */ void (*close)(struct dpif *dpif); /* Attempts to destroy the dpif underlying 'dpif'. * * If successful, 'dpif' will not be used again except as an argument for * the 'close' member function. */ int (*destroy)(struct dpif *dpif); /* Performs periodic work needed by 'dpif', if any is necessary. * Returns true if need to revalidate. */ bool (*run)(struct dpif *dpif); /* Arranges for poll_block() to wake up if the "run" member function needs * to be called for 'dpif'. */ void (*wait)(struct dpif *dpif); /* Retrieves statistics for 'dpif' into 'stats'. */ int (*get_stats)(const struct dpif *dpif, struct dpif_dp_stats *stats); int (*set_features)(struct dpif *dpif, uint32_t user_features); /* Adds 'netdev' as a new port in 'dpif'. If '*port_no' is not * ODPP_NONE, attempts to use that as the port's port number. * * If port is successfully added, sets '*port_no' to the new port's * port number. Returns EBUSY if caller attempted to choose a port * number, and it was in use. */ int (*port_add)(struct dpif *dpif, struct netdev *netdev, odp_port_t *port_no); /* Removes port numbered 'port_no' from 'dpif'. */ int (*port_del)(struct dpif *dpif, odp_port_t port_no); /* Refreshes configuration of 'dpif's port. The implementation might * postpone applying the changes until run() is called. */ int (*port_set_config)(struct dpif *dpif, odp_port_t port_no, const struct smap *cfg); /* Queries 'dpif' for a port with the given 'port_no' or 'devname'. * If 'port' is not null, stores information about the port into * '*port' if successful. * * If the port doesn't exist, the provider must return ENODEV. Other * error numbers means that something wrong happened and will be * treated differently by upper layers. * * If 'port' is not null, the caller takes ownership of data in * 'port' and must free it with dpif_port_destroy() when it is no * longer needed. */ int (*port_query_by_number)(const struct dpif *dpif, odp_port_t port_no, struct dpif_port *port); int (*port_query_by_name)(const struct dpif *dpif, const char *devname, struct dpif_port *port); /* Returns the Netlink PID value to supply in OVS_ACTION_ATTR_USERSPACE * actions as the OVS_USERSPACE_ATTR_PID attribute's value, for use in * flows whose packets arrived on port 'port_no'. * * A 'port_no' of UINT32_MAX should be treated as a special case. The * implementation should return a reserved PID, not allocated to any port, * that the client may use for special purposes. * * The return value only needs to be meaningful when DPIF_UC_ACTION has * been enabled in the 'dpif''s listen mask, and it is allowed to change * when DPIF_UC_ACTION is disabled and then re-enabled. * * A dpif provider that doesn't have meaningful Netlink PIDs can use NULL * for this function. This is equivalent to always returning 0. */ uint32_t (*port_get_pid)(const struct dpif *dpif, odp_port_t port_no); /* Attempts to begin dumping the ports in a dpif. On success, returns 0 * and initializes '*statep' with any data needed for iteration. On * failure, returns a positive errno value. */ int (*port_dump_start)(const struct dpif *dpif, void **statep); /* Attempts to retrieve another port from 'dpif' for 'state', which was * initialized by a successful call to the 'port_dump_start' function for * 'dpif'. On success, stores a new dpif_port into 'port' and returns 0. * Returns EOF if the end of the port table has been reached, or a positive * errno value on error. This function will not be called again once it * returns nonzero once for a given iteration (but the 'port_dump_done' * function will be called afterward). * * The dpif provider retains ownership of the data stored in 'port'. It * must remain valid until at least the next call to 'port_dump_next' or * 'port_dump_done' for 'state'. */ int (*port_dump_next)(const struct dpif *dpif, void *state, struct dpif_port *port); /* Releases resources from 'dpif' for 'state', which was initialized by a * successful call to the 'port_dump_start' function for 'dpif'. */ int (*port_dump_done)(const struct dpif *dpif, void *state); /* Polls for changes in the set of ports in 'dpif'. If the set of ports in * 'dpif' has changed, then this function should do one of the * following: * * - Preferably: store the name of the device that was added to or deleted * from 'dpif' in '*devnamep' and return 0. The caller is responsible * for freeing '*devnamep' (with free()) when it no longer needs it. * * - Alternatively: return ENOBUFS, without indicating the device that was * added or deleted. * * Occasional 'false positives', in which the function returns 0 while * indicating a device that was not actually added or deleted or returns * ENOBUFS without any change, are acceptable. * * If the set of ports in 'dpif' has not changed, returns EAGAIN. May also * return other positive errno values to indicate that something has gone * wrong. */ int (*port_poll)(const struct dpif *dpif, char **devnamep); /* Arranges for the poll loop to wake up when 'port_poll' will return a * value other than EAGAIN. */ void (*port_poll_wait)(const struct dpif *dpif); /* Deletes all flows from 'dpif' and clears all of its queues of received * packets. */ int (*flow_flush)(struct dpif *dpif); /* Flow dumping interface. * * This is the back-end for the flow dumping interface described in * dpif.h. Please read the comments there first, because this code * closely follows it. * * 'flow_dump_create' and 'flow_dump_thread_create' must always return an * initialized and usable data structure and defer error return until * flow_dump_destroy(). This hasn't been a problem for the dpifs that * exist so far. * * 'flow_dump_create' and 'flow_dump_thread_create' must initialize the * structures that they return with dpif_flow_dump_init() and * dpif_flow_dump_thread_init(), respectively. * * If 'terse' is true, then only UID and statistics will * be returned in the dump. Otherwise, all fields will be returned. * * If 'types' isn't null, dumps only the flows of the passed types. */ struct dpif_flow_dump *(*flow_dump_create)( const struct dpif *dpif, bool terse, struct dpif_flow_dump_types *types); int (*flow_dump_destroy)(struct dpif_flow_dump *dump); struct dpif_flow_dump_thread *(*flow_dump_thread_create)( struct dpif_flow_dump *dump); void (*flow_dump_thread_destroy)(struct dpif_flow_dump_thread *thread); int (*flow_dump_next)(struct dpif_flow_dump_thread *thread, struct dpif_flow *flows, int max_flows); /* Executes each of the 'n_ops' operations in 'ops' on 'dpif', in the order * in which they are specified, placing each operation's results in the * "output" members documented in comments and the 'error' member of each * dpif_op. The offload_type argument tells the provider if 'ops' should * be submitted to to a netdev (only offload) or to the kernel datapath * (never offload) or to both (offload if possible; software fallback). */ void (*operate)(struct dpif *dpif, struct dpif_op **ops, size_t n_ops, enum dpif_offload_type offload_type); /* Get hardware-offloads activity counters from a dataplane. * Those counters are not offload statistics (which are accessible through * netdev statistics), but a status of hardware offload management: * how many offloads are currently waiting, inserted, etc. */ int (*offload_stats_get)(struct dpif *dpif, struct netdev_custom_stats *stats); /* Enables or disables receiving packets with dpif_recv() for 'dpif'. * Turning packet receive off and then back on is allowed to change Netlink * PID assignments (see ->port_get_pid()). The client is responsible for * updating flows as necessary if it does this. */ int (*recv_set)(struct dpif *dpif, bool enable); /* Attempts to refresh the poll loops and Netlink sockets used for handling * upcalls when the number of upcall handlers (upcall receiving thread) is * changed to 'n_handlers' and receiving packets for 'dpif' is enabled by * recv_set(). * * A dpif implementation may choose to ignore 'n_handlers' while returning * success. * * The method for distribution of upcalls between handler threads is * specific to the dpif implementation. */ int (*handlers_set)(struct dpif *dpif, uint32_t n_handlers); /* Queries 'dpif' to see if a certain number of handlers are required by * the implementation. * * If a certain number of handlers are required, returns 'true' and sets * 'n_handlers' to that number of handler threads. * * If not, returns 'false'. */ bool (*number_handlers_required)(struct dpif *dpif, uint32_t *n_handlers); /* Pass custom configuration options to the datapath. The implementation * might postpone applying the changes until run() is called. */ int (*set_config)(struct dpif *dpif, const struct smap *other_config); /* Translates OpenFlow queue ID 'queue_id' (in host byte order) into a * priority value used for setting packet priority. */ int (*queue_to_priority)(const struct dpif *dpif, uint32_t queue_id, uint32_t *priority); /* Polls for an upcall from 'dpif' for an upcall handler. Since there * can be multiple poll loops (see ->handlers_set()), 'handler_id' is * needed as index to identify the corresponding poll loop. If * successful, stores the upcall into '*upcall', using 'buf' for * storage. Should only be called if 'recv_set' has been used to enable * receiving packets from 'dpif'. * * The implementation should point 'upcall->key' and 'upcall->userdata' * (if any) into data in the caller-provided 'buf'. The implementation may * also use 'buf' for storing the data of 'upcall->packet'. If necessary * to make room, the implementation may reallocate the data in 'buf'. * * The caller owns the data of 'upcall->packet' and may modify it. If * packet's headroom is exhausted as it is manipulated, 'upcall->packet' * will be reallocated. This requires the data of 'upcall->packet' to be * released with ofpbuf_uninit() before 'upcall' is destroyed. However, * when an error is returned, the 'upcall->packet' may be uninitialized * and should not be released. * * This function must not block. If no upcall is pending when it is * called, it should return EAGAIN without blocking. */ int (*recv)(struct dpif *dpif, uint32_t handler_id, struct dpif_upcall *upcall, struct ofpbuf *buf); /* Arranges for the poll loop for an upcall handler to wake up when 'dpif' * has a message queued to be received with the recv member functions. * Since there can be multiple poll loops (see ->handlers_set()), * 'handler_id' is needed as index to identify the corresponding poll loop. * */ void (*recv_wait)(struct dpif *dpif, uint32_t handler_id); /* Throws away any queued upcalls that 'dpif' currently has ready to * return. */ void (*recv_purge)(struct dpif *dpif); /* When 'dpif' is about to purge the datapath, the higher layer may want * to be notified so that it could try reacting accordingly (e.g. grabbing * all flow stats before they are gone). * * Registers an upcall callback function with 'dpif'. This is only used * if 'dpif' needs to notify the purging of datapath. 'aux' is passed to * the callback on invocation. */ void (*register_dp_purge_cb)(struct dpif *, dp_purge_callback *, void *aux); /* For datapaths that run in userspace (i.e. dpif-netdev), threads polling * for incoming packets can directly call upcall functions instead of * offloading packet processing to separate handler threads. Datapaths * that directly call upcall functions should use the functions below to * to register an upcall function and enable / disable upcalls. * * Registers an upcall callback function with 'dpif'. This is only used * if 'dpif' directly executes upcall functions. 'aux' is passed to the * callback on invocation. */ void (*register_upcall_cb)(struct dpif *, upcall_callback *, void *aux); /* Enables upcalls if 'dpif' directly executes upcall functions. */ void (*enable_upcall)(struct dpif *); /* Disables upcalls if 'dpif' directly executes upcall functions. */ void (*disable_upcall)(struct dpif *); /* Get datapath version. Caller is responsible for freeing the string * returned. */ char *(*get_datapath_version)(void); /* Conntrack entry dumping interface. * * These functions are used by ct-dpif.c to provide a datapath-agnostic * dumping interface to the connection trackers provided by the * datapaths. * * ct_dump_start() should put in '*state' a pointer to a newly allocated * stucture that will be passed by the caller to ct_dump_next() and * ct_dump_done(). If 'zone' is not NULL, only the entries in '*zone' * should be dumped. * * ct_dump_next() should fill 'entry' with information from a connection * and prepare to dump the next one on a subsequest invocation. * * ct_dump_done() should perform any cleanup necessary (including * deallocating the 'state' structure, if applicable). */ int (*ct_dump_start)(struct dpif *, struct ct_dpif_dump_state **state, const uint16_t *zone, int *); int (*ct_dump_next)(struct dpif *, struct ct_dpif_dump_state *state, struct ct_dpif_entry *entry); int (*ct_dump_done)(struct dpif *, struct ct_dpif_dump_state *state); /* Flushes the connection tracking tables. The arguments have the * following behavior: * * - If both 'zone' and 'tuple' are NULL, flush all the conntrack * entries. * - If 'zone' is not NULL, and 'tuple' is NULL, flush all the * conntrack entries in '*zone'. * - If 'tuple' is not NULL, flush the conntrack entry specified by * 'tuple' in '*zone'. If 'zone' is NULL, use the default zone * (zone 0). */ int (*ct_flush)(struct dpif *, const uint16_t *zone, const struct ct_dpif_tuple *tuple); /* Set max connections allowed. */ int (*ct_set_maxconns)(struct dpif *, uint32_t maxconns); /* Get max connections allowed. */ int (*ct_get_maxconns)(struct dpif *, uint32_t *maxconns); /* Get number of connections tracked. */ int (*ct_get_nconns)(struct dpif *, uint32_t *nconns); /* Enable or disable TCP sequence checking. */ int (*ct_set_tcp_seq_chk)(struct dpif *, bool enabled); /* Get the TCP sequence checking configuration. */ int (*ct_get_tcp_seq_chk)(struct dpif *, bool *enabled); /* Updates the sweep interval for the CT sweeper. */ int (*ct_set_sweep_interval)(struct dpif *, uint32_t ms); /* Get the current value of the sweep interval for the CT sweeper. */ int (*ct_get_sweep_interval)(struct dpif *, uint32_t *ms); /* Connection tracking per zone limit */ /* Per zone conntrack limit sets the maximum allowed connections in zones * to provide resource isolation. If a per zone limit for a particular * zone is not available in the datapath, it defaults to the default * per zone limit. Initially, the default per zone limit is * unlimited (0). */ /* Sets the max connections allowed per zone according to 'zone_limits', * a list of 'struct ct_dpif_zone_limit' entries (the 'count' member * is not used when setting limits). If 'default_limit' is not NULL, * modifies the default limit to '*default_limit'. */ int (*ct_set_limits)(struct dpif *, const uint32_t *default_limit, const struct ovs_list *zone_limits); /* Looks up the default per zone limit and stores that in * 'default_limit'. Look up the per zone limits for all zones in * the 'zone_limits_in' list of 'struct ct_dpif_zone_limit' entries * (the 'limit' and 'count' members are not used), and stores the * reply that includes the zone, the per zone limit, and the number * of connections in the zone into 'zone_limits_out' list. */ int (*ct_get_limits)(struct dpif *, uint32_t *default_limit, const struct ovs_list *zone_limits_in, struct ovs_list *zone_limits_out); /* Deletes per zone limit of all zones specified in 'zone_limits', a * list of 'struct ct_dpif_zone_limit' entries. */ int (*ct_del_limits)(struct dpif *, const struct ovs_list *zone_limits); /* Connection tracking timeout policy */ /* A connection tracking timeout policy contains a list of timeout * attributes that specify timeout values on various connection states. * In a datapath, the timeout policy is identified by a 4-byte unsigned * integer. Unsupported timeout attributes are ignored. When a * connection is committed it can be associated with a timeout * policy, or it defaults to the datapath's default timeout policy. */ /* Sets timeout policy '*tp' into the datapath. */ int (*ct_set_timeout_policy)(struct dpif *, const struct ct_dpif_timeout_policy *tp); /* Gets a timeout policy specified by tp_id and stores it into '*tp'. */ int (*ct_get_timeout_policy)(struct dpif *, uint32_t tp_id, struct ct_dpif_timeout_policy *tp); /* Deletes a timeout policy identified by 'tp_id'. */ int (*ct_del_timeout_policy)(struct dpif *, uint32_t tp_id); /* Conntrack timeout policy dumping interface. * * These functions provide a datapath-agnostic dumping interface * to the conntrack timeout policy provided by the datapaths. * * ct_timeout_policy_dump_start() should put in '*statep' a pointer to * a newly allocated structure that will be passed by the caller to * ct_timeout_policy_dump_next() and ct_timeout_policy_dump_done(). * * ct_timeout_policy_dump_next() attempts to retrieve another timeout * policy from 'dpif' for 'state', which was initialized by a successful * call to ct_timeout_policy_dump_start(). On success, stores a new * timeout policy into 'tp' and returns 0. Returns EOF if the last * timeout policy has been dumped, or a positive errno value on error. * This function will not be called again once it returns nonzero once * for a given iteration (but the ct_timeout_policy_dump_done() will * be called afterward). * * ct_timeout_policy_dump_done() should perform any cleanup necessary * (including deallocating the 'state' structure, if applicable). */ int (*ct_timeout_policy_dump_start)(struct dpif *, void **statep); int (*ct_timeout_policy_dump_next)(struct dpif *, void *state, struct ct_dpif_timeout_policy *tp); int (*ct_timeout_policy_dump_done)(struct dpif *, void *state); /* Gets timeout policy based on 'tp_id', 'dl_type' and 'nw_proto'. * On success, returns 0, stores the timeout policy name in 'tp_name', * and sets 'is_generic'. 'is_generic' is false if the returned timeout * policy in the 'dpif' is specific to 'dl_type' and 'nw_proto' in the * datapath (e.g., the Linux kernel datapath). Sets 'is_generic' to * true, if the timeout policy supports all OVS supported L3/L4 * protocols. * * The caller is responsible for freeing 'tp_name'. */ int (*ct_get_timeout_policy_name)(struct dpif *, uint32_t tp_id, uint16_t dl_type, uint8_t nw_proto, char **tp_name, bool *is_generic); /* Stores the conntrack features supported by 'dpif' into features. * The value is a bitmap of CONNTRACK_F_* bits. */ int (*ct_get_features)(struct dpif *, enum ct_features *features); /* IP Fragmentation. */ /* Disables or enables conntrack fragment reassembly. The default * setting is enabled. */ int (*ipf_set_enabled)(struct dpif *, bool v6, bool enabled); /* Set minimum fragment allowed. */ int (*ipf_set_min_frag)(struct dpif *, bool v6, uint32_t min_frag); /* Set maximum number of fragments tracked. */ int (*ipf_set_max_nfrags)(struct dpif *, uint32_t max_nfrags); /* Get fragmentation configuration status and counters. */ int (*ipf_get_status)(struct dpif *, struct dpif_ipf_status *dpif_ipf_status); /* The following 3 apis find and print ipf lists by creating a string * representation of the state of an ipf list, to which 'dump' is pointed * to. 'ipf_dump_start()' allocates memory for 'ipf_dump_ctx'. * 'ipf_dump_next()' finds the next ipf list and copies it's * characteristics to a string, which is freed by the caller. * 'ipf_dump_done()' frees the 'ipf_dump_ctx' that was allocated in * 'ipf_dump_start'. */ int (*ipf_dump_start)(struct dpif *, struct ipf_dump_ctx **ipf_dump_ctx); int (*ipf_dump_next)(struct dpif *, void *ipf_dump_ctx, char **dump); int (*ipf_dump_done)(struct dpif *, void *ipf_dump_ctx); /* Meters */ /* Queries 'dpif' for supported meter features. * NULL pointer means no meter features are supported. */ void (*meter_get_features)(const struct dpif *, struct ofputil_meter_features *); /* Adds or modifies the meter in 'dpif' with the given 'meter_id' * and the configuration in 'config'. * * The meter id specified through 'config->meter_id' is ignored. */ int (*meter_set)(struct dpif *, ofproto_meter_id meter_id, struct ofputil_meter_config *); /* Queries 'dpif' for meter stats with the given 'meter_id'. Stores * maximum of 'n_bands' meter statistics, returning the number of band * stats returned in 'stats->n_bands' if successful. */ int (*meter_get)(const struct dpif *, ofproto_meter_id meter_id, struct ofputil_meter_stats *, uint16_t n_bands); /* Removes meter 'meter_id' from 'dpif'. Stores meter and band statistics * (for maximum of 'n_bands', returning the number of band stats returned * in 'stats->n_bands' if successful. 'stats' may be passed in as NULL if * no stats are needed, in which case 'n_bands' must be passed in as * zero. */ int (*meter_del)(struct dpif *, ofproto_meter_id meter_id, struct ofputil_meter_stats *, uint16_t n_bands); /* Adds a bond with 'bond_id' and the member-map to 'dpif'. */ int (*bond_add)(struct dpif *dpif, uint32_t bond_id, odp_port_t *member_map); /* Removes bond identified by 'bond_id' from 'dpif'. */ int (*bond_del)(struct dpif *dpif, uint32_t bond_id); /* Reads bond stats from 'dpif'. 'n_bytes' should be an array with size * sufficient to store BOND_BUCKETS number of elements. */ int (*bond_stats_get)(struct dpif *dpif, uint32_t bond_id, uint64_t *n_bytes); /* Cache configuration * * Multiple levels of cache can exist in a given datapath implementation. * An API has been provided to get the number of supported caches, which * can then be used to get/set specific configuration. Cache level is 0 * indexed, i.e. if 1 level is supported, the level value to use is 0. * * Get the number of cache levels supported. */ int (*cache_get_supported_levels)(struct dpif *dpif, uint32_t *levels); /* Get the cache name for the given level. */ int (*cache_get_name)(struct dpif *dpif, uint32_t level, const char **name); /* Get currently configured cache size. */ int (*cache_get_size)(struct dpif *dpif, uint32_t level, uint32_t *size); /* Set cache size. */ int (*cache_set_size)(struct dpif *dpif, uint32_t level, uint32_t size); }; extern const struct dpif_class dpif_netlink_class; extern const struct dpif_class dpif_netdev_class; #ifdef __cplusplus } #endif #endif /* dpif-provider.h */