diff mbox

[rdma-core,02/10] Remove uses of htonll

Message ID 1486677212-3188-3-git-send-email-jgunthorpe@obsidianresearch.com (mailing list archive)
State Accepted
Headers show

Commit Message

Jason Gunthorpe Feb. 9, 2017, 9:53 p.m. UTC
Done via

$ git ls-files | xargs sed -i -e 's/htonll/htobe64/g'

Signed-off-by: Jason Gunthorpe <jgunthorpe@obsidianresearch.com>
---
 ibacm/include/acm_mad.h       | 84 +++++++++++++++++++++----------------------
 ibacm/prov/acmp/src/acmp.c    | 10 +++---
 ibacm/src/acm.c               |  6 ++--
 providers/cxgb3/cxio_wr.h     |  2 +-
 providers/cxgb3/qp.c          | 12 +++----
 providers/cxgb4/t4.h          |  2 +-
 providers/mlx4/qp.c           | 12 +++----
 providers/mlx4/srq.c          |  2 +-
 providers/mlx5/qp.c           | 24 ++++++-------
 providers/mlx5/srq.c          |  2 +-
 providers/mthca/qp.c          | 34 +++++++++---------
 providers/mthca/srq.c         |  4 +--
 providers/qedr/qelr_verbs.c   |  2 +-
 srp_daemon/srp_daemon.c       | 16 ++++-----
 srp_daemon/srp_handle_traps.c |  4 +--
 15 files changed, 108 insertions(+), 108 deletions(-)

Comments

Steve Wise Feb. 9, 2017, 10:23 p.m. UTC | #1
> Subject: [PATCH rdma-core 02/10] Remove uses of htonll
> 
> Done via
> 
> $ git ls-files | xargs sed -i -e 's/htonll/htobe64/g'
> 
> Signed-off-by: Jason Gunthorpe <jgunthorpe@obsidianresearch.com>
> ---
>  ibacm/include/acm_mad.h       | 84
+++++++++++++++++++++----------------------
>  ibacm/prov/acmp/src/acmp.c    | 10 +++---
>  ibacm/src/acm.c               |  6 ++--
>  providers/cxgb3/cxio_wr.h     |  2 +-
>  providers/cxgb3/qp.c          | 12 +++----
>  providers/cxgb4/t4.h          |  2 +-
>  providers/mlx4/qp.c           | 12 +++----
>  providers/mlx4/srq.c          |  2 +-
>  providers/mlx5/qp.c           | 24 ++++++-------
>  providers/mlx5/srq.c          |  2 +-
>  providers/mthca/qp.c          | 34 +++++++++---------
>  providers/mthca/srq.c         |  4 +--
>  providers/qedr/qelr_verbs.c   |  2 +-
>  srp_daemon/srp_daemon.c       | 16 ++++-----
>  srp_daemon/srp_handle_traps.c |  4 +--
>  15 files changed, 108 insertions(+), 108 deletions(-)

cxgb* parts look fine.

Reviewed-by: Steve Wise <swise@opengridcomputing.com>


--
To unsubscribe from this list: send the line "unsubscribe linux-rdma" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Amrani, Ram Feb. 12, 2017, 8:26 a.m. UTC | #2
> 
> Done via
> 
> $ git ls-files | xargs sed -i -e 's/htonll/htobe64/g'
> 
> Signed-off-by: Jason Gunthorpe <jgunthorpe@obsidianresearch.com>
> ---
>  ibacm/include/acm_mad.h       | 84 +++++++++++++++++++++----------------------
>  ibacm/prov/acmp/src/acmp.c    | 10 +++---
>  ibacm/src/acm.c               |  6 ++--
>  providers/cxgb3/cxio_wr.h     |  2 +-
>  providers/cxgb3/qp.c          | 12 +++----
>  providers/cxgb4/t4.h          |  2 +-
>  providers/mlx4/qp.c           | 12 +++----
>  providers/mlx4/srq.c          |  2 +-
>  providers/mlx5/qp.c           | 24 ++++++-------
>  providers/mlx5/srq.c          |  2 +-
>  providers/mthca/qp.c          | 34 +++++++++---------
>  providers/mthca/srq.c         |  4 +--
>  providers/qedr/qelr_verbs.c   |  2 +-
>  srp_daemon/srp_daemon.c       | 16 ++++-----
>  srp_daemon/srp_handle_traps.c |  4 +--
>  15 files changed, 108 insertions(+), 108 deletions(-)
> 
> diff --git a/ibacm/include/acm_mad.h b/ibacm/include/acm_mad.h
> index 68e231bc6680ca..f237ef6fde563c 100644
> --- a/ibacm/include/acm_mad.h
> +++ b/ibacm/include/acm_mad.h
> @@ -121,52 +121,52 @@ struct ib_sa_mad {
> 
>  #define IB_SA_ATTR_PATH_REC htons(0x0035)
> 
> -#define IB_COMP_MASK_PR_SERVICE_ID         (htonll(1 << 0) | \
> -                                            htonll(1 << 1))
> -#define IB_COMP_MASK_PR_DGID                htonll(1 << 2)
> -#define IB_COMP_MASK_PR_SGID                htonll(1 << 3)
> -#define IB_COMP_MASK_PR_DLID                htonll(1 << 4)
> -#define IB_COMP_MASK_PR_SLID                htonll(1 << 5)
> -#define IB_COMP_MASK_PR_RAW_TRAFFIC         htonll(1 << 6)
> -/* RESERVED                                 htonll(1 << 7) */
> -#define IB_COMP_MASK_PR_FLOW_LABEL          htonll(1 << 8)
> -#define IB_COMP_MASK_PR_HOP_LIMIT           htonll(1 << 9)
> -#define IB_COMP_MASK_PR_TCLASS              htonll(1 << 10)
> -#define IB_COMP_MASK_PR_REVERSIBLE          htonll(1 << 11)
> -#define IB_COMP_MASK_PR_NUM_PATH            htonll(1 << 12)
> -#define IB_COMP_MASK_PR_PKEY                htonll(1 << 13)
> -#define IB_COMP_MASK_PR_QOS_CLASS           htonll(1 << 14)
> -#define IB_COMP_MASK_PR_SL                  htonll(1 << 15)
> -#define IB_COMP_MASK_PR_MTU_SELECTOR        htonll(1 << 16)
> -#define IB_COMP_MASK_PR_MTU                 htonll(1 << 17)
> -#define IB_COMP_MASK_PR_RATE_SELECTOR       htonll(1 << 18)
> -#define IB_COMP_MASK_PR_RATE                htonll(1 << 19)
> -#define IB_COMP_MASK_PR_PACKET_LIFETIME_SELECTOR htonll(1 << 20)
> -#define IB_COMP_MASK_PR_PACKET_LIFETIME     htonll(1 << 21)
> -#define IB_COMP_MASK_PR_PREFERENCE          htonll(1 << 22)
> -/* RESERVED                                 htonll(1 << 23) */
> +#define IB_COMP_MASK_PR_SERVICE_ID         (htobe64(1 << 0) | \
> +                                            htobe64(1 << 1))
> +#define IB_COMP_MASK_PR_DGID                htobe64(1 << 2)
> +#define IB_COMP_MASK_PR_SGID                htobe64(1 << 3)
> +#define IB_COMP_MASK_PR_DLID                htobe64(1 << 4)
> +#define IB_COMP_MASK_PR_SLID                htobe64(1 << 5)
> +#define IB_COMP_MASK_PR_RAW_TRAFFIC         htobe64(1 << 6)
> +/* RESERVED                                 htobe64(1 << 7) */
> +#define IB_COMP_MASK_PR_FLOW_LABEL          htobe64(1 << 8)
> +#define IB_COMP_MASK_PR_HOP_LIMIT           htobe64(1 << 9)
> +#define IB_COMP_MASK_PR_TCLASS              htobe64(1 << 10)
> +#define IB_COMP_MASK_PR_REVERSIBLE          htobe64(1 << 11)
> +#define IB_COMP_MASK_PR_NUM_PATH            htobe64(1 << 12)
> +#define IB_COMP_MASK_PR_PKEY                htobe64(1 << 13)
> +#define IB_COMP_MASK_PR_QOS_CLASS           htobe64(1 << 14)
> +#define IB_COMP_MASK_PR_SL                  htobe64(1 << 15)
> +#define IB_COMP_MASK_PR_MTU_SELECTOR        htobe64(1 << 16)
> +#define IB_COMP_MASK_PR_MTU                 htobe64(1 << 17)
> +#define IB_COMP_MASK_PR_RATE_SELECTOR       htobe64(1 << 18)
> +#define IB_COMP_MASK_PR_RATE                htobe64(1 << 19)
> +#define IB_COMP_MASK_PR_PACKET_LIFETIME_SELECTOR htobe64(1 << 20)
> +#define IB_COMP_MASK_PR_PACKET_LIFETIME     htobe64(1 << 21)
> +#define IB_COMP_MASK_PR_PREFERENCE          htobe64(1 << 22)
> +/* RESERVED                                 htobe64(1 << 23) */
> 
>  #define IB_MC_QPN 0xffffff
>  #define IB_SA_ATTR_MC_MEMBER_REC htons(0x0038)
> 
> -#define IB_COMP_MASK_MC_MGID                htonll(1 << 0)
> -#define IB_COMP_MASK_MC_PORT_GID            htonll(1 << 1)
> -#define IB_COMP_MASK_MC_QKEY                htonll(1 << 2)
> -#define IB_COMP_MASK_MC_MLID                htonll(1 << 3)
> -#define IB_COMP_MASK_MC_MTU_SEL             htonll(1 << 4)
> -#define IB_COMP_MASK_MC_MTU                 htonll(1 << 5)
> -#define IB_COMP_MASK_MC_TCLASS              htonll(1 << 6)
> -#define IB_COMP_MASK_MC_PKEY                htonll(1 << 7)
> -#define IB_COMP_MASK_MC_RATE_SEL            htonll(1 << 8)
> -#define IB_COMP_MASK_MC_RATE                htonll(1 << 9)
> -#define IB_COMP_MASK_MC_PACKET_LIFETIME_SEL htonll(1 << 10)
> -#define IB_COMP_MASK_MC_PACKET_LIFETIME     htonll(1 << 11)
> -#define IB_COMP_MASK_MC_SL                  htonll(1 << 12)
> -#define IB_COMP_MASK_MC_FLOW                htonll(1 << 13)
> -#define IB_COMP_MASK_MC_HOP                 htonll(1 << 14)
> -#define IB_COMP_MASK_MC_SCOPE               htonll(1 << 15)
> -#define IB_COMP_MASK_MC_JOIN_STATE          htonll(1 << 16)
> -#define IB_COMP_MASK_MC_PROXY_JOIN          htonll(1 << 17)
> +#define IB_COMP_MASK_MC_MGID                htobe64(1 << 0)
> +#define IB_COMP_MASK_MC_PORT_GID            htobe64(1 << 1)
> +#define IB_COMP_MASK_MC_QKEY                htobe64(1 << 2)
> +#define IB_COMP_MASK_MC_MLID                htobe64(1 << 3)
> +#define IB_COMP_MASK_MC_MTU_SEL             htobe64(1 << 4)
> +#define IB_COMP_MASK_MC_MTU                 htobe64(1 << 5)
> +#define IB_COMP_MASK_MC_TCLASS              htobe64(1 << 6)
> +#define IB_COMP_MASK_MC_PKEY                htobe64(1 << 7)
> +#define IB_COMP_MASK_MC_RATE_SEL            htobe64(1 << 8)
> +#define IB_COMP_MASK_MC_RATE                htobe64(1 << 9)
> +#define IB_COMP_MASK_MC_PACKET_LIFETIME_SEL htobe64(1 << 10)
> +#define IB_COMP_MASK_MC_PACKET_LIFETIME     htobe64(1 << 11)
> +#define IB_COMP_MASK_MC_SL                  htobe64(1 << 12)
> +#define IB_COMP_MASK_MC_FLOW                htobe64(1 << 13)
> +#define IB_COMP_MASK_MC_HOP                 htobe64(1 << 14)
> +#define IB_COMP_MASK_MC_SCOPE               htobe64(1 << 15)
> +#define IB_COMP_MASK_MC_JOIN_STATE          htobe64(1 << 16)
> +#define IB_COMP_MASK_MC_PROXY_JOIN          htobe64(1 << 17)
> 
>  struct ib_mc_member_rec {
>  	union ibv_gid mgid;
> diff --git a/ibacm/prov/acmp/src/acmp.c b/ibacm/prov/acmp/src/acmp.c
> index fb6684f2eac1f5..6b7cbb8c203478 100644
> --- a/ibacm/prov/acmp/src/acmp.c
> +++ b/ibacm/prov/acmp/src/acmp.c
> @@ -780,7 +780,7 @@ static void acmp_init_path_query(struct ib_sa_mad *mad)
>  	mad->mgmt_class = IB_MGMT_CLASS_SA;
>  	mad->class_version = 2;
>  	mad->method = IB_METHOD_GET;
> -	mad->tid = htonll((uint64_t) atomic_inc(&g_tid));
> +	mad->tid = htobe64((uint64_t) atomic_inc(&g_tid));
>  	mad->attr_id = IB_SA_ATTR_PATH_REC;
>  }
> 
> @@ -1367,7 +1367,7 @@ static void acmp_init_join(struct ib_sa_mad *mad, union ibv_gid *port_gid,
>  	mad->mgmt_class = IB_MGMT_CLASS_SA;
>  	mad->class_version = 2;
>  	mad->method = IB_METHOD_SET;
> -	mad->tid = htonll((uint64_t) atomic_inc(&g_tid));
> +	mad->tid = htobe64((uint64_t) atomic_inc(&g_tid));
>  	mad->attr_id = IB_SA_ATTR_MC_MEMBER_REC;
>  	mad->comp_mask =
>  		IB_COMP_MASK_MC_MGID | IB_COMP_MASK_MC_PORT_GID |
> @@ -1673,7 +1673,7 @@ acmp_send_resolve(struct acmp_ep *ep, struct acmp_dest *dest,
>  	mad->class_version = 1;
>  	mad->method = IB_METHOD_GET;
>  	mad->control = ACM_CTRL_RESOLVE;
> -	mad->tid = htonll((uint64_t) atomic_inc(&g_tid));
> +	mad->tid = htobe64((uint64_t) atomic_inc(&g_tid));
> 
>  	rec = (struct acm_resolve_rec *) mad->data;
>  	rec->src_type = (uint8_t) saddr->type;
> @@ -1957,7 +1957,7 @@ static void acmp_query_perf(void *ep_context, uint64_t *values, uint8_t *cnt)
>  	int i;
> 
>  	for (i = 0; i < ACM_MAX_COUNTER; i++)
> -		values[i] = htonll((uint64_t) atomic_get(&ep->counters[i]));
> +		values[i] = htobe64((uint64_t) atomic_get(&ep->counters[i]));
>  	*cnt = ACM_MAX_COUNTER;
>  }
> 
> @@ -2082,7 +2082,7 @@ static void acmp_parse_osm_fullv1_lid2guid(FILE *f, uint64_t *lid2guid)
>  		if (lid2guid[lid])
>  			acm_log(0, "ERROR - duplicate lid %u\n", lid);
>  		else
> -			lid2guid[lid] = htonll(guid);
> +			lid2guid[lid] = htobe64(guid);
>  	}
>  }
> 
> diff --git a/ibacm/src/acm.c b/ibacm/src/acm.c
> index fd06a2e81fbe76..2cf1ea8b2148c0 100644
> --- a/ibacm/src/acm.c
> +++ b/ibacm/src/acm.c
> @@ -1000,7 +1000,7 @@ static int acm_svr_perf_query(struct acmc_client *client, struct acm_msg *msg)
>  	    ((ntohs(msg->hdr.length) >= (ACM_MSG_HDR_LENGTH + ACM_MSG_EP_LENGTH)
>  	    && !(msg->resolve_data[0].flags & ACM_EP_FLAG_SOURCE)))) {
>  		for (i = 0; i < ACM_MAX_COUNTER; i++)
> -			msg->perf_data[i] = htonll((uint64_t) atomic_get(&counter[i]));
> +			msg->perf_data[i] = htobe64((uint64_t) atomic_get(&counter[i]));
> 
>  		msg->hdr.data[0] = ACM_MAX_COUNTER;
>  		len = ACM_MSG_HDR_LENGTH + (ACM_MAX_COUNTER * sizeof(uint64_t));
> @@ -1426,7 +1426,7 @@ static int acm_nl_parse_path_attr(struct nlattr *attr,
>  		sid = (uint64_t *) NLA_DATA(attr);
>  		if (NLA_LEN(attr) == sizeof(*sid)) {
>  			acm_log(2, "service_id 0x%" PRIx64 "\n", *sid);
> -			path->service_id = htonll(*sid);
> +			path->service_id = htobe64(*sid);
>  		} else {
>  			ret = -1;
>  		}
> @@ -2499,7 +2499,7 @@ static void acm_load_prov_config(void)
>  		acm_log(2, "provider %s subnet_prefix 0x%" PRIx64 "\n",
>  			prov_name, prefix);
>  		/* Convert it into network byte order */
> -		prefix = htonll(prefix);
> +		prefix = htobe64(prefix);
> 
>  		list_for_each(&provider_list, prov, entry) {
>  			if (!strcasecmp(prov->prov->name, prov_name)) {
> diff --git a/providers/cxgb3/cxio_wr.h b/providers/cxgb3/cxio_wr.h
> index d56c5298ebaad5..735b64918a15c8 100644
> --- a/providers/cxgb3/cxio_wr.h
> +++ b/providers/cxgb3/cxio_wr.h
> @@ -351,7 +351,7 @@ static inline void build_fw_riwrh(struct fw_riwrh *wqe, enum t3_wr_opcode op,
>  	wqe->gen_tid_len = htonl(V_FW_RIWR_GEN(genbit) | V_FW_RIWR_TID(tid) |
>  				 V_FW_RIWR_LEN(len));
>  	/* 2nd gen bit... */
> -        ((union t3_wr *)wqe)->flit[15] = htonll(genbit);
> +        ((union t3_wr *)wqe)->flit[15] = htobe64(genbit);
>  }
> 
>  /*
> diff --git a/providers/cxgb3/qp.c b/providers/cxgb3/qp.c
> index 30dd898f6e874b..b0cf10364e9a6c 100644
> --- a/providers/cxgb3/qp.c
> +++ b/providers/cxgb3/qp.c
> @@ -84,7 +84,7 @@ static inline int iwch_build_rdma_send(union t3_wr *wqe, struct ibv_send_wr *wr,
>  			    htonl(wr->sg_list[i].lkey);
>  			wqe->send.sgl[i].len =
>  			    htonl(wr->sg_list[i].length);
> -			wqe->send.sgl[i].to = htonll(wr->sg_list[i].addr);
> +			wqe->send.sgl[i].to = htobe64(wr->sg_list[i].addr);
>  		}
>  		wqe->send.plen = htonl(wqe->send.plen);
>  		wqe->send.num_sgle = htonl(wr->num_sge);
> @@ -104,7 +104,7 @@ static inline int iwch_build_rdma_write(union t3_wr *wqe,
>  	wqe->write.rdmaop = T3_RDMA_WRITE;
>  	wqe->write.reserved = 0;
>  	wqe->write.stag_sink = htonl(wr->wr.rdma.rkey);
> -	wqe->write.to_sink = htonll(wr->wr.rdma.remote_addr);
> +	wqe->write.to_sink = htobe64(wr->wr.rdma.remote_addr);
> 
>  	wqe->write.num_sgle = wr->num_sge;
> 
> @@ -139,7 +139,7 @@ static inline int iwch_build_rdma_write(union t3_wr *wqe,
>  			wqe->write.sgl[i].len =
>  			    htonl(wr->sg_list[i].length);
>  			wqe->write.sgl[i].to =
> -			    htonll(wr->sg_list[i].addr);
> +			    htobe64(wr->sg_list[i].addr);
>  		}
>  		wqe->write.plen = htonl(wqe->write.plen);
>  		wqe->write.num_sgle = htonl(wr->num_sge);
> @@ -157,10 +157,10 @@ static inline int iwch_build_rdma_read(union t3_wr *wqe, struct ibv_send_wr *wr,
>  	wqe->read.reserved = 0;
>  	if (wr->num_sge == 1 && wr->sg_list[0].length > 0) {
>  		wqe->read.rem_stag = htonl(wr->wr.rdma.rkey);
> -		wqe->read.rem_to = htonll(wr->wr.rdma.remote_addr);
> +		wqe->read.rem_to = htobe64(wr->wr.rdma.remote_addr);
>  		wqe->read.local_stag = htonl(wr->sg_list[0].lkey);
>  		wqe->read.local_len = htonl(wr->sg_list[0].length);
> -		wqe->read.local_to = htonll(wr->sg_list[0].addr);
> +		wqe->read.local_to = htobe64(wr->sg_list[0].addr);
>  	} else {
> 
>  		/* build passable 0B read request */
> @@ -295,7 +295,7 @@ static inline int iwch_build_rdma_recv(struct iwch_device *rhp,
>  	for (i = 0; i < wr->num_sge; i++) {
>  		wqe->recv.sgl[i].stag = htonl(wr->sg_list[i].lkey);
>  		wqe->recv.sgl[i].len = htonl(wr->sg_list[i].length);
> -		wqe->recv.sgl[i].to = htonll(wr->sg_list[i].addr);
> +		wqe->recv.sgl[i].to = htobe64(wr->sg_list[i].addr);
>  	}
>  	for (; i < T3_MAX_SGE; i++) {
>  		wqe->recv.sgl[i].stag = 0;
> diff --git a/providers/cxgb4/t4.h b/providers/cxgb4/t4.h
> index cad675c3fdb193..59dbc65d4e164b 100644
> --- a/providers/cxgb4/t4.h
> +++ b/providers/cxgb4/t4.h
> @@ -55,7 +55,7 @@
>  #define __iomem
>  #define cpu_to_be16 htons
>  #define cpu_to_be32 htonl
> -#define cpu_to_be64 htonll
> +#define cpu_to_be64 htobe64
>  #define be16_to_cpu ntohs
>  #define be32_to_cpu ntohl
>  #define be64_to_cpu ntohll
> diff --git a/providers/mlx4/qp.c b/providers/mlx4/qp.c
> index 268fb7dc83dd11..a607326c7c452c 100644
> --- a/providers/mlx4/qp.c
> +++ b/providers/mlx4/qp.c
> @@ -157,7 +157,7 @@ static inline void set_local_inv_seg(struct mlx4_wqe_local_inval_seg *iseg,
>  static inline void set_raddr_seg(struct mlx4_wqe_raddr_seg *rseg,
>  				 uint64_t remote_addr, uint32_t rkey)
>  {
> -	rseg->raddr    = htonll(remote_addr);
> +	rseg->raddr    = htobe64(remote_addr);
>  	rseg->rkey     = htonl(rkey);
>  	rseg->reserved = 0;
>  }
> @@ -165,10 +165,10 @@ static inline void set_raddr_seg(struct mlx4_wqe_raddr_seg *rseg,
>  static void set_atomic_seg(struct mlx4_wqe_atomic_seg *aseg, struct ibv_send_wr *wr)
>  {
>  	if (wr->opcode == IBV_WR_ATOMIC_CMP_AND_SWP) {
> -		aseg->swap_add = htonll(wr->wr.atomic.swap);
> -		aseg->compare  = htonll(wr->wr.atomic.compare_add);
> +		aseg->swap_add = htobe64(wr->wr.atomic.swap);
> +		aseg->compare  = htobe64(wr->wr.atomic.compare_add);
>  	} else {
> -		aseg->swap_add = htonll(wr->wr.atomic.compare_add);
> +		aseg->swap_add = htobe64(wr->wr.atomic.compare_add);
>  		aseg->compare  = 0;
>  	}
> 
> @@ -188,13 +188,13 @@ static void __set_data_seg(struct mlx4_wqe_data_seg *dseg, struct ibv_sge *sg)
>  {
>  	dseg->byte_count = htonl(sg->length);
>  	dseg->lkey       = htonl(sg->lkey);
> -	dseg->addr       = htonll(sg->addr);
> +	dseg->addr       = htobe64(sg->addr);
>  }
> 
>  static void set_data_seg(struct mlx4_wqe_data_seg *dseg, struct ibv_sge *sg)
>  {
>  	dseg->lkey       = htonl(sg->lkey);
> -	dseg->addr       = htonll(sg->addr);
> +	dseg->addr       = htobe64(sg->addr);
> 
>  	/*
>  	 * Need a barrier here before writing the byte_count field to
> diff --git a/providers/mlx4/srq.c b/providers/mlx4/srq.c
> index c0e028671828d8..4f90efdf927209 100644
> --- a/providers/mlx4/srq.c
> +++ b/providers/mlx4/srq.c
> @@ -96,7 +96,7 @@ int mlx4_post_srq_recv(struct ibv_srq *ibsrq,
>  		for (i = 0; i < wr->num_sge; ++i) {
>  			scat[i].byte_count = htonl(wr->sg_list[i].length);
>  			scat[i].lkey       = htonl(wr->sg_list[i].lkey);
> -			scat[i].addr       = htonll(wr->sg_list[i].addr);
> +			scat[i].addr       = htobe64(wr->sg_list[i].addr);
>  		}
> 
>  		if (i < srq->max_gs) {
> diff --git a/providers/mlx5/qp.c b/providers/mlx5/qp.c
> index e82b1a0bebc3f9..7d034880f4caf6 100644
> --- a/providers/mlx5/qp.c
> +++ b/providers/mlx5/qp.c
> @@ -190,7 +190,7 @@ static int mlx5_wq_overflow(struct mlx5_wq *wq, int nreq, struct mlx5_cq *cq)
>  static inline void set_raddr_seg(struct mlx5_wqe_raddr_seg *rseg,
>  				 uint64_t remote_addr, uint32_t rkey)
>  {
> -	rseg->raddr    = htonll(remote_addr);
> +	rseg->raddr    = htobe64(remote_addr);
>  	rseg->rkey     = htonl(rkey);
>  	rseg->reserved = 0;
>  }
> @@ -201,10 +201,10 @@ static void set_atomic_seg(struct mlx5_wqe_atomic_seg *aseg,
>  			   uint64_t compare_add)
>  {
>  	if (opcode == IBV_WR_ATOMIC_CMP_AND_SWP) {
> -		aseg->swap_add = htonll(swap);
> -		aseg->compare  = htonll(compare_add);
> +		aseg->swap_add = htobe64(swap);
> +		aseg->compare  = htobe64(compare_add);
>  	} else {
> -		aseg->swap_add = htonll(compare_add);
> +		aseg->swap_add = htobe64(compare_add);
>  	}
>  }
> 
> @@ -221,7 +221,7 @@ static void set_data_ptr_seg(struct mlx5_wqe_data_seg *dseg, struct ibv_sge *sg,
>  {
>  	dseg->byte_count = htonl(sg->length - offset);
>  	dseg->lkey       = htonl(sg->lkey);
> -	dseg->addr       = htonll(sg->addr + offset);
> +	dseg->addr       = htobe64(sg->addr + offset);
>  }
> 
>  static void set_data_ptr_seg_atomic(struct mlx5_wqe_data_seg *dseg,
> @@ -229,7 +229,7 @@ static void set_data_ptr_seg_atomic(struct mlx5_wqe_data_seg *dseg,
>  {
>  	dseg->byte_count = htonl(MLX5_ATOMIC_SIZE);
>  	dseg->lkey       = htonl(sg->lkey);
> -	dseg->addr       = htonll(sg->addr);
> +	dseg->addr       = htobe64(sg->addr);
>  }
> 
>  /*
> @@ -430,7 +430,7 @@ static void set_umr_data_seg(struct mlx5_qp *qp, enum ibv_mw_type type,
> 
>  	data->klm.byte_count = htonl(bind_info->length);
>  	data->klm.mkey = htonl(bind_info->mr->lkey);
> -	data->klm.address = htonll(bind_info->addr);
> +	data->klm.address = htobe64(bind_info->addr);
> 
>  	memset(&data->klm + 1, 0, sizeof(data->reserved) -
>  	       sizeof(data->klm));
> @@ -467,8 +467,8 @@ static void set_umr_mkey_seg(struct mlx5_qp *qp, enum ibv_mw_type type,
>  		if (bind_info->mw_access_flags & IBV_ACCESS_ZERO_BASED)
>  			mkey->start_addr = 0;
>  		else
> -			mkey->start_addr = htonll(bind_info->addr);
> -		mkey->len = htonll(bind_info->length);
> +			mkey->start_addr = htobe64(bind_info->addr);
> +		mkey->len = htobe64(bind_info->length);
>  	} else {
>  		mkey->free = MLX5_WQE_MKEY_CONTEXT_FREE;
>  	}
> @@ -485,20 +485,20 @@ static inline void set_umr_control_seg(struct mlx5_qp *qp, enum ibv_mw_type type
> 
>  	ctrl->flags = MLX5_WQE_UMR_CTRL_FLAG_TRNSLATION_OFFSET |
>  		MLX5_WQE_UMR_CTRL_FLAG_INLINE;
> -	ctrl->mkey_mask = htonll(MLX5_WQE_UMR_CTRL_MKEY_MASK_FREE |
> +	ctrl->mkey_mask = htobe64(MLX5_WQE_UMR_CTRL_MKEY_MASK_FREE |
>  				     MLX5_WQE_UMR_CTRL_MKEY_MASK_MKEY);
>  	ctrl->translation_offset = 0;
>  	memset(ctrl->rsvd0, 0, sizeof(ctrl->rsvd0));
>  	memset(ctrl->rsvd1, 0, sizeof(ctrl->rsvd1));
> 
>  	if (type == IBV_MW_TYPE_2)
> -		ctrl->mkey_mask |= htonll(MLX5_WQE_UMR_CTRL_MKEY_MASK_QPN);
> +		ctrl->mkey_mask |= htobe64(MLX5_WQE_UMR_CTRL_MKEY_MASK_QPN);
> 
>  	if (bind_info->length) {
>  		ctrl->klm_octowords = get_klm_octo(1);
>  		if (type == IBV_MW_TYPE_2)
>  			ctrl->flags |=  MLX5_WQE_UMR_CTRL_FLAG_CHECK_FREE;
> -		ctrl->mkey_mask |= htonll(MLX5_WQE_UMR_CTRL_MKEY_MASK_LEN	|
> +		ctrl->mkey_mask |= htobe64(MLX5_WQE_UMR_CTRL_MKEY_MASK_LEN	|
>  					      MLX5_WQE_UMR_CTRL_MKEY_MASK_START_ADDR |
>  					      MLX5_WQE_UMR_CTRL_MKEY_MASK_ACCESS_LOCAL_WRITE |
>  					      MLX5_WQE_UMR_CTRL_MKEY_MASK_ACCESS_REMOTE_READ |
> diff --git a/providers/mlx5/srq.c b/providers/mlx5/srq.c
> index a06afa3a5931b6..b362ec85724961 100644
> --- a/providers/mlx5/srq.c
> +++ b/providers/mlx5/srq.c
> @@ -120,7 +120,7 @@ int mlx5_post_srq_recv(struct ibv_srq *ibsrq,
>  		for (i = 0; i < wr->num_sge; ++i) {
>  			scat[i].byte_count = htonl(wr->sg_list[i].length);
>  			scat[i].lkey       = htonl(wr->sg_list[i].lkey);
> -			scat[i].addr       = htonll(wr->sg_list[i].addr);
> +			scat[i].addr       = htobe64(wr->sg_list[i].addr);
>  		}
> 
>  		if (i < srq->max_gs) {
> diff --git a/providers/mthca/qp.c b/providers/mthca/qp.c
> index 129e6c74ef5382..d221bb19bfa67c 100644
> --- a/providers/mthca/qp.c
> +++ b/providers/mthca/qp.c
> @@ -147,7 +147,7 @@ int mthca_tavor_post_send(struct ibv_qp *ibqp, struct ibv_send_wr *wr,
>  			case IBV_WR_ATOMIC_CMP_AND_SWP:
>  			case IBV_WR_ATOMIC_FETCH_AND_ADD:
>  				((struct mthca_raddr_seg *) wqe)->raddr =
> -					htonll(wr->wr.atomic.remote_addr);
> +					htobe64(wr->wr.atomic.remote_addr);
>  				((struct mthca_raddr_seg *) wqe)->rkey =
>  					htonl(wr->wr.atomic.rkey);
>  				((struct mthca_raddr_seg *) wqe)->reserved = 0;
> @@ -156,12 +156,12 @@ int mthca_tavor_post_send(struct ibv_qp *ibqp, struct ibv_send_wr *wr,
> 
>  				if (wr->opcode == IBV_WR_ATOMIC_CMP_AND_SWP) {
>  					((struct mthca_atomic_seg *) wqe)->swap_add =
> -						htonll(wr->wr.atomic.swap);
> +						htobe64(wr->wr.atomic.swap);
>  					((struct mthca_atomic_seg *) wqe)->compare =
> -						htonll(wr->wr.atomic.compare_add);
> +						htobe64(wr->wr.atomic.compare_add);
>  				} else {
>  					((struct mthca_atomic_seg *) wqe)->swap_add =
> -						htonll(wr->wr.atomic.compare_add);
> +						htobe64(wr->wr.atomic.compare_add);
>  					((struct mthca_atomic_seg *) wqe)->compare = 0;
>  				}
> 
> @@ -174,7 +174,7 @@ int mthca_tavor_post_send(struct ibv_qp *ibqp, struct ibv_send_wr *wr,
>  			case IBV_WR_RDMA_WRITE_WITH_IMM:
>  			case IBV_WR_RDMA_READ:
>  				((struct mthca_raddr_seg *) wqe)->raddr =
> -					htonll(wr->wr.rdma.remote_addr);
> +					htobe64(wr->wr.rdma.remote_addr);
>  				((struct mthca_raddr_seg *) wqe)->rkey =
>  					htonl(wr->wr.rdma.rkey);
>  				((struct mthca_raddr_seg *) wqe)->reserved = 0;
> @@ -194,7 +194,7 @@ int mthca_tavor_post_send(struct ibv_qp *ibqp, struct ibv_send_wr *wr,
>  			case IBV_WR_RDMA_WRITE:
>  			case IBV_WR_RDMA_WRITE_WITH_IMM:
>  				((struct mthca_raddr_seg *) wqe)->raddr =
> -					htonll(wr->wr.rdma.remote_addr);
> +					htobe64(wr->wr.rdma.remote_addr);
>  				((struct mthca_raddr_seg *) wqe)->rkey =
>  					htonl(wr->wr.rdma.rkey);
>  				((struct mthca_raddr_seg *) wqe)->reserved = 0;
> @@ -213,7 +213,7 @@ int mthca_tavor_post_send(struct ibv_qp *ibqp, struct ibv_send_wr *wr,
>  			((struct mthca_tavor_ud_seg *) wqe)->lkey =
>  				htonl(to_mah(wr->wr.ud.ah)->key);
>  			((struct mthca_tavor_ud_seg *) wqe)->av_addr =
> -				htonll((uintptr_t) to_mah(wr->wr.ud.ah)->av);
> +				htobe64((uintptr_t) to_mah(wr->wr.ud.ah)->av);
>  			((struct mthca_tavor_ud_seg *) wqe)->dqpn =
>  				htonl(wr->wr.ud.remote_qpn);
>  			((struct mthca_tavor_ud_seg *) wqe)->qkey =
> @@ -265,7 +265,7 @@ int mthca_tavor_post_send(struct ibv_qp *ibqp, struct ibv_send_wr *wr,
>  				seg = wqe;
>  				seg->byte_count = htonl(wr->sg_list[i].length);
>  				seg->lkey = htonl(wr->sg_list[i].lkey);
> -				seg->addr = htonll(wr->sg_list[i].addr);
> +				seg->addr = htobe64(wr->sg_list[i].addr);
>  				wqe += sizeof *seg;
>  			}
> 
> @@ -372,7 +372,7 @@ int mthca_tavor_post_recv(struct ibv_qp *ibqp, struct ibv_recv_wr *wr,
>  			((struct mthca_data_seg *) wqe)->lkey =
>  				htonl(wr->sg_list[i].lkey);
>  			((struct mthca_data_seg *) wqe)->addr =
> -				htonll(wr->sg_list[i].addr);
> +				htobe64(wr->sg_list[i].addr);
>  			wqe += sizeof (struct mthca_data_seg);
>  			size += sizeof (struct mthca_data_seg) / 16;
>  		}
> @@ -508,7 +508,7 @@ int mthca_arbel_post_send(struct ibv_qp *ibqp, struct ibv_send_wr *wr,
>  			case IBV_WR_ATOMIC_CMP_AND_SWP:
>  			case IBV_WR_ATOMIC_FETCH_AND_ADD:
>  				((struct mthca_raddr_seg *) wqe)->raddr =
> -					htonll(wr->wr.atomic.remote_addr);
> +					htobe64(wr->wr.atomic.remote_addr);
>  				((struct mthca_raddr_seg *) wqe)->rkey =
>  					htonl(wr->wr.atomic.rkey);
>  				((struct mthca_raddr_seg *) wqe)->reserved = 0;
> @@ -517,12 +517,12 @@ int mthca_arbel_post_send(struct ibv_qp *ibqp, struct ibv_send_wr *wr,
> 
>  				if (wr->opcode == IBV_WR_ATOMIC_CMP_AND_SWP) {
>  					((struct mthca_atomic_seg *) wqe)->swap_add =
> -						htonll(wr->wr.atomic.swap);
> +						htobe64(wr->wr.atomic.swap);
>  					((struct mthca_atomic_seg *) wqe)->compare =
> -						htonll(wr->wr.atomic.compare_add);
> +						htobe64(wr->wr.atomic.compare_add);
>  				} else {
>  					((struct mthca_atomic_seg *) wqe)->swap_add =
> -						htonll(wr->wr.atomic.compare_add);
> +						htobe64(wr->wr.atomic.compare_add);
>  					((struct mthca_atomic_seg *) wqe)->compare = 0;
>  				}
> 
> @@ -535,7 +535,7 @@ int mthca_arbel_post_send(struct ibv_qp *ibqp, struct ibv_send_wr *wr,
>  			case IBV_WR_RDMA_WRITE_WITH_IMM:
>  			case IBV_WR_RDMA_READ:
>  				((struct mthca_raddr_seg *) wqe)->raddr =
> -					htonll(wr->wr.rdma.remote_addr);
> +					htobe64(wr->wr.rdma.remote_addr);
>  				((struct mthca_raddr_seg *) wqe)->rkey =
>  					htonl(wr->wr.rdma.rkey);
>  				((struct mthca_raddr_seg *) wqe)->reserved = 0;
> @@ -555,7 +555,7 @@ int mthca_arbel_post_send(struct ibv_qp *ibqp, struct ibv_send_wr *wr,
>  			case IBV_WR_RDMA_WRITE:
>  			case IBV_WR_RDMA_WRITE_WITH_IMM:
>  				((struct mthca_raddr_seg *) wqe)->raddr =
> -					htonll(wr->wr.rdma.remote_addr);
> +					htobe64(wr->wr.rdma.remote_addr);
>  				((struct mthca_raddr_seg *) wqe)->rkey =
>  					htonl(wr->wr.rdma.rkey);
>  				((struct mthca_raddr_seg *) wqe)->reserved = 0;
> @@ -624,7 +624,7 @@ int mthca_arbel_post_send(struct ibv_qp *ibqp, struct ibv_send_wr *wr,
>  				seg = wqe;
>  				seg->byte_count = htonl(wr->sg_list[i].length);
>  				seg->lkey = htonl(wr->sg_list[i].lkey);
> -				seg->addr = htonll(wr->sg_list[i].addr);
> +				seg->addr = htobe64(wr->sg_list[i].addr);
>  				wqe += sizeof *seg;
>  			}
> 
> @@ -730,7 +730,7 @@ int mthca_arbel_post_recv(struct ibv_qp *ibqp, struct ibv_recv_wr *wr,
>  			((struct mthca_data_seg *) wqe)->lkey =
>  				htonl(wr->sg_list[i].lkey);
>  			((struct mthca_data_seg *) wqe)->addr =
> -				htonll(wr->sg_list[i].addr);
> +				htobe64(wr->sg_list[i].addr);
>  			wqe += sizeof (struct mthca_data_seg);
>  		}
> 
> diff --git a/providers/mthca/srq.c b/providers/mthca/srq.c
> index ff6f532733f4ae..66ac924a720c84 100644
> --- a/providers/mthca/srq.c
> +++ b/providers/mthca/srq.c
> @@ -126,7 +126,7 @@ int mthca_tavor_post_srq_recv(struct ibv_srq *ibsrq,
>  			((struct mthca_data_seg *) wqe)->lkey =
>  				htonl(wr->sg_list[i].lkey);
>  			((struct mthca_data_seg *) wqe)->addr =
> -				htonll(wr->sg_list[i].addr);
> +				htobe64(wr->sg_list[i].addr);
>  			wqe += sizeof (struct mthca_data_seg);
>  		}
> 
> @@ -219,7 +219,7 @@ int mthca_arbel_post_srq_recv(struct ibv_srq *ibsrq,
>  			((struct mthca_data_seg *) wqe)->lkey =
>  				htonl(wr->sg_list[i].lkey);
>  			((struct mthca_data_seg *) wqe)->addr =
> -				htonll(wr->sg_list[i].addr);
> +				htobe64(wr->sg_list[i].addr);
>  			wqe += sizeof (struct mthca_data_seg);
>  		}
> 
> diff --git a/providers/qedr/qelr_verbs.c b/providers/qedr/qelr_verbs.c
> index 823ed8754c09d7..932115c8ab7f39 100644
> --- a/providers/qedr/qelr_verbs.c
> +++ b/providers/qedr/qelr_verbs.c
> @@ -900,7 +900,7 @@ static inline void qelr_edpm_set_rdma_ext(struct qelr_qp *qp,
>  	if (!qp->edpm.is_edpm)
>  		return;
> 
> -	qp->edpm.rdma_ext->remote_va = htonll(remote_addr);
> +	qp->edpm.rdma_ext->remote_va = htobe64(remote_addr);
>  	qp->edpm.rdma_ext->remote_key = htonl(rkey);
>  	qp->edpm.dpm_payload_offset += sizeof(*qp->edpm.rdma_ext);
>  	qp->edpm.dpm_payload_size += sizeof(*qp->edpm.rdma_ext);
> diff --git a/srp_daemon/srp_daemon.c b/srp_daemon/srp_daemon.c
> index e0e550d81ab3c2..ff6118acf611ca 100644
> --- a/srp_daemon/srp_daemon.c
> +++ b/srp_daemon/srp_daemon.c
> @@ -462,9 +462,9 @@ static int add_non_exist_target(struct target_details *target)
>  			if (srpd_sys_read_gid(scsi_host_dir, "dgid", dgid_val))
>  				continue;
>  		}
> -		if (htonll(target->subnet_prefix) != *((uint64_t *) dgid_val))
> +		if (htobe64(target->subnet_prefix) != *((uint64_t *) dgid_val))
>  			continue;
> -		if (htonll(target->h_guid) != *((uint64_t *) (dgid_val+8)))
> +		if (htobe64(target->h_guid) != *((uint64_t *) (dgid_val+8)))
>  			continue;
> 
>  		/* If there is no local_ib_device in the scsi host dir (old kernel module), assumes it is equal */
> @@ -603,7 +603,7 @@ static int send_and_get(int portid, int agent, srp_ib_user_mad_t *out_mad,
>  		/* Skip tid 0 because OpenSM ignores it. */
>  		if (++tid == 0)
>  			++tid;
> -		out_dm_mad->tid = htonll(tid);
> +		out_dm_mad->tid = htobe64(tid);
> 
>  		ret = umad_send(portid, agent, out_mad, MAD_BLOCK_SIZE,
>  				config->timeout, 0);
> @@ -1033,7 +1033,7 @@ int get_node(struct umad_resources *umad_res, uint16_t dlid, uint64_t *guid)
>  	init_srp_sa_mad(&out_mad, umad_res->agent, umad_res->sm_lid,
>  		        SRP_SA_ATTR_NODE, 0);
> 
> -	out_sa_mad->comp_mask     = htonll(1); /* LID */
> +	out_sa_mad->comp_mask     = htobe64(1); /* LID */
>  	node			  = (void *) out_sa_mad->data;
>  	node->lid		  = htons(dlid);
> 
> @@ -1059,7 +1059,7 @@ static int get_port_info(struct umad_resources *umad_res, uint16_t dlid,
>  	init_srp_sa_mad(&out_mad, umad_res->agent, umad_res->sm_lid,
>  		        SRP_SA_ATTR_PORT_INFO, 0);
> 
> -	out_sa_mad->comp_mask     = htonll(1); /* LID */
> +	out_sa_mad->comp_mask     = htobe64(1); /* LID */
>  	port_info                 = (void *) out_sa_mad->data;
>  	port_info->endport_lid	  = htons(dlid);
> 
> @@ -1134,7 +1134,7 @@ static int get_shared_pkeys(struct resources *res,
>  			continue;
> 
>  		/* Mark components: DLID, SLID, PKEY */
> -		out_sa_mad->comp_mask = htonll(1 << 4 | 1 << 5 | 1 << 13);
> +		out_sa_mad->comp_mask = htobe64(1 << 4 | 1 << 5 | 1 << 13);
>  		out_sa_mad->rmpp_version = 1;
>  		out_sa_mad->rmpp_type = 1;
>  		path_rec = (ib_path_rec_t *)out_sa_mad->data;
> @@ -1194,7 +1194,7 @@ static int do_dm_port_list(struct resources *res)
>  		        SRP_SA_ATTR_PORT_INFO, SRP_SM_CAP_MASK_MATCH_ATTR_MOD);
> 
>  	out_sa_mad->method     	   = SRP_SA_METHOD_GET_TABLE;
> -	out_sa_mad->comp_mask      = htonll(1 << 7); /* Capability mask */
> +	out_sa_mad->comp_mask      = htobe64(1 << 7); /* Capability mask */
>  	out_sa_mad->rmpp_version   = 1;
>  	out_sa_mad->rmpp_type      = 1;
>  	port_info		   = (void *) out_sa_mad->data;
> @@ -2330,7 +2330,7 @@ static int get_lid(struct umad_resources *umad_res, ib_gid_t *gid, uint16_t *lid
>  	init_srp_sa_mad(&out_mad, umad_res->agent, umad_res->sm_lid,
>  		        SRP_SA_ATTR_PATH_REC, 0);
> 
> -	out_sa_mad->comp_mask = htonll( 4 | 8 | 64 | 512 | 4096 );
> +	out_sa_mad->comp_mask = htobe64( 4 | 8 | 64 | 512 | 4096 );
> 
>  	path_rec->sgid = *gid;
>  	path_rec->dgid = *gid;
> diff --git a/srp_daemon/srp_handle_traps.c b/srp_daemon/srp_handle_traps.c
> index 01cc5445e8352b..6a7a2d1ab479a1 100644
> --- a/srp_daemon/srp_handle_traps.c
> +++ b/srp_daemon/srp_handle_traps.c
> @@ -606,14 +606,14 @@ static int register_to_trap(struct sync_resources *sync_res,
>  	    comp_mask |= SRP_INFORMINFO_QPN_COMP;
>  	}
> 
> -	p_sa_mad->comp_mask = htonll(comp_mask);
> +	p_sa_mad->comp_mask = htobe64(comp_mask);
>  	pr_debug("comp_mask: %llx\n", comp_mask);
> 
>  	do {
>  		pthread_mutex_lock(res->mad_buffer_mutex);
>  		res->mad_buffer->base_ver = 0; // flag that the buffer is empty
>  		pthread_mutex_unlock(res->mad_buffer_mutex);
> -		mad_hdr->trans_id = htonll(trans_id);
> +		mad_hdr->trans_id = htobe64(trans_id);
>  		trans_id++;
> 
>  		ret = ibv_post_send(res->qp, &sr, bad_wr);
> --
> 2.7.4
> 

Looks good.

Reviewed-by: Ram Amrani <Ram.Amrani@cavium.com>

--
To unsubscribe from this list: send the line "unsubscribe linux-rdma" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
diff mbox

Patch

diff --git a/ibacm/include/acm_mad.h b/ibacm/include/acm_mad.h
index 68e231bc6680ca..f237ef6fde563c 100644
--- a/ibacm/include/acm_mad.h
+++ b/ibacm/include/acm_mad.h
@@ -121,52 +121,52 @@  struct ib_sa_mad {
 
 #define IB_SA_ATTR_PATH_REC htons(0x0035)
 
-#define IB_COMP_MASK_PR_SERVICE_ID         (htonll(1 << 0) | \
-                                            htonll(1 << 1))
-#define IB_COMP_MASK_PR_DGID                htonll(1 << 2)
-#define IB_COMP_MASK_PR_SGID                htonll(1 << 3)
-#define IB_COMP_MASK_PR_DLID                htonll(1 << 4)
-#define IB_COMP_MASK_PR_SLID                htonll(1 << 5)
-#define IB_COMP_MASK_PR_RAW_TRAFFIC         htonll(1 << 6)
-/* RESERVED                                 htonll(1 << 7) */
-#define IB_COMP_MASK_PR_FLOW_LABEL          htonll(1 << 8)
-#define IB_COMP_MASK_PR_HOP_LIMIT           htonll(1 << 9)
-#define IB_COMP_MASK_PR_TCLASS              htonll(1 << 10)
-#define IB_COMP_MASK_PR_REVERSIBLE          htonll(1 << 11)
-#define IB_COMP_MASK_PR_NUM_PATH            htonll(1 << 12)
-#define IB_COMP_MASK_PR_PKEY                htonll(1 << 13)
-#define IB_COMP_MASK_PR_QOS_CLASS           htonll(1 << 14)
-#define IB_COMP_MASK_PR_SL                  htonll(1 << 15)
-#define IB_COMP_MASK_PR_MTU_SELECTOR        htonll(1 << 16)
-#define IB_COMP_MASK_PR_MTU                 htonll(1 << 17)
-#define IB_COMP_MASK_PR_RATE_SELECTOR       htonll(1 << 18)
-#define IB_COMP_MASK_PR_RATE                htonll(1 << 19)
-#define IB_COMP_MASK_PR_PACKET_LIFETIME_SELECTOR htonll(1 << 20)
-#define IB_COMP_MASK_PR_PACKET_LIFETIME     htonll(1 << 21)
-#define IB_COMP_MASK_PR_PREFERENCE          htonll(1 << 22)
-/* RESERVED                                 htonll(1 << 23) */
+#define IB_COMP_MASK_PR_SERVICE_ID         (htobe64(1 << 0) | \
+                                            htobe64(1 << 1))
+#define IB_COMP_MASK_PR_DGID                htobe64(1 << 2)
+#define IB_COMP_MASK_PR_SGID                htobe64(1 << 3)
+#define IB_COMP_MASK_PR_DLID                htobe64(1 << 4)
+#define IB_COMP_MASK_PR_SLID                htobe64(1 << 5)
+#define IB_COMP_MASK_PR_RAW_TRAFFIC         htobe64(1 << 6)
+/* RESERVED                                 htobe64(1 << 7) */
+#define IB_COMP_MASK_PR_FLOW_LABEL          htobe64(1 << 8)
+#define IB_COMP_MASK_PR_HOP_LIMIT           htobe64(1 << 9)
+#define IB_COMP_MASK_PR_TCLASS              htobe64(1 << 10)
+#define IB_COMP_MASK_PR_REVERSIBLE          htobe64(1 << 11)
+#define IB_COMP_MASK_PR_NUM_PATH            htobe64(1 << 12)
+#define IB_COMP_MASK_PR_PKEY                htobe64(1 << 13)
+#define IB_COMP_MASK_PR_QOS_CLASS           htobe64(1 << 14)
+#define IB_COMP_MASK_PR_SL                  htobe64(1 << 15)
+#define IB_COMP_MASK_PR_MTU_SELECTOR        htobe64(1 << 16)
+#define IB_COMP_MASK_PR_MTU                 htobe64(1 << 17)
+#define IB_COMP_MASK_PR_RATE_SELECTOR       htobe64(1 << 18)
+#define IB_COMP_MASK_PR_RATE                htobe64(1 << 19)
+#define IB_COMP_MASK_PR_PACKET_LIFETIME_SELECTOR htobe64(1 << 20)
+#define IB_COMP_MASK_PR_PACKET_LIFETIME     htobe64(1 << 21)
+#define IB_COMP_MASK_PR_PREFERENCE          htobe64(1 << 22)
+/* RESERVED                                 htobe64(1 << 23) */
 
 #define IB_MC_QPN 0xffffff
 #define IB_SA_ATTR_MC_MEMBER_REC htons(0x0038)
 
-#define IB_COMP_MASK_MC_MGID                htonll(1 << 0)
-#define IB_COMP_MASK_MC_PORT_GID            htonll(1 << 1)
-#define IB_COMP_MASK_MC_QKEY                htonll(1 << 2)
-#define IB_COMP_MASK_MC_MLID                htonll(1 << 3)
-#define IB_COMP_MASK_MC_MTU_SEL             htonll(1 << 4)
-#define IB_COMP_MASK_MC_MTU                 htonll(1 << 5)
-#define IB_COMP_MASK_MC_TCLASS              htonll(1 << 6)
-#define IB_COMP_MASK_MC_PKEY                htonll(1 << 7)
-#define IB_COMP_MASK_MC_RATE_SEL            htonll(1 << 8)
-#define IB_COMP_MASK_MC_RATE                htonll(1 << 9)
-#define IB_COMP_MASK_MC_PACKET_LIFETIME_SEL htonll(1 << 10)
-#define IB_COMP_MASK_MC_PACKET_LIFETIME     htonll(1 << 11)
-#define IB_COMP_MASK_MC_SL                  htonll(1 << 12)
-#define IB_COMP_MASK_MC_FLOW                htonll(1 << 13)
-#define IB_COMP_MASK_MC_HOP                 htonll(1 << 14)
-#define IB_COMP_MASK_MC_SCOPE               htonll(1 << 15)
-#define IB_COMP_MASK_MC_JOIN_STATE          htonll(1 << 16)
-#define IB_COMP_MASK_MC_PROXY_JOIN          htonll(1 << 17)
+#define IB_COMP_MASK_MC_MGID                htobe64(1 << 0)
+#define IB_COMP_MASK_MC_PORT_GID            htobe64(1 << 1)
+#define IB_COMP_MASK_MC_QKEY                htobe64(1 << 2)
+#define IB_COMP_MASK_MC_MLID                htobe64(1 << 3)
+#define IB_COMP_MASK_MC_MTU_SEL             htobe64(1 << 4)
+#define IB_COMP_MASK_MC_MTU                 htobe64(1 << 5)
+#define IB_COMP_MASK_MC_TCLASS              htobe64(1 << 6)
+#define IB_COMP_MASK_MC_PKEY                htobe64(1 << 7)
+#define IB_COMP_MASK_MC_RATE_SEL            htobe64(1 << 8)
+#define IB_COMP_MASK_MC_RATE                htobe64(1 << 9)
+#define IB_COMP_MASK_MC_PACKET_LIFETIME_SEL htobe64(1 << 10)
+#define IB_COMP_MASK_MC_PACKET_LIFETIME     htobe64(1 << 11)
+#define IB_COMP_MASK_MC_SL                  htobe64(1 << 12)
+#define IB_COMP_MASK_MC_FLOW                htobe64(1 << 13)
+#define IB_COMP_MASK_MC_HOP                 htobe64(1 << 14)
+#define IB_COMP_MASK_MC_SCOPE               htobe64(1 << 15)
+#define IB_COMP_MASK_MC_JOIN_STATE          htobe64(1 << 16)
+#define IB_COMP_MASK_MC_PROXY_JOIN          htobe64(1 << 17)
 
 struct ib_mc_member_rec {
 	union ibv_gid mgid;
diff --git a/ibacm/prov/acmp/src/acmp.c b/ibacm/prov/acmp/src/acmp.c
index fb6684f2eac1f5..6b7cbb8c203478 100644
--- a/ibacm/prov/acmp/src/acmp.c
+++ b/ibacm/prov/acmp/src/acmp.c
@@ -780,7 +780,7 @@  static void acmp_init_path_query(struct ib_sa_mad *mad)
 	mad->mgmt_class = IB_MGMT_CLASS_SA;
 	mad->class_version = 2;
 	mad->method = IB_METHOD_GET;
-	mad->tid = htonll((uint64_t) atomic_inc(&g_tid));
+	mad->tid = htobe64((uint64_t) atomic_inc(&g_tid));
 	mad->attr_id = IB_SA_ATTR_PATH_REC;
 }
 
@@ -1367,7 +1367,7 @@  static void acmp_init_join(struct ib_sa_mad *mad, union ibv_gid *port_gid,
 	mad->mgmt_class = IB_MGMT_CLASS_SA;
 	mad->class_version = 2;
 	mad->method = IB_METHOD_SET;
-	mad->tid = htonll((uint64_t) atomic_inc(&g_tid));
+	mad->tid = htobe64((uint64_t) atomic_inc(&g_tid));
 	mad->attr_id = IB_SA_ATTR_MC_MEMBER_REC;
 	mad->comp_mask =
 		IB_COMP_MASK_MC_MGID | IB_COMP_MASK_MC_PORT_GID |
@@ -1673,7 +1673,7 @@  acmp_send_resolve(struct acmp_ep *ep, struct acmp_dest *dest,
 	mad->class_version = 1;
 	mad->method = IB_METHOD_GET;
 	mad->control = ACM_CTRL_RESOLVE;
-	mad->tid = htonll((uint64_t) atomic_inc(&g_tid));
+	mad->tid = htobe64((uint64_t) atomic_inc(&g_tid));
 
 	rec = (struct acm_resolve_rec *) mad->data;
 	rec->src_type = (uint8_t) saddr->type;
@@ -1957,7 +1957,7 @@  static void acmp_query_perf(void *ep_context, uint64_t *values, uint8_t *cnt)
 	int i;
 
 	for (i = 0; i < ACM_MAX_COUNTER; i++)
-		values[i] = htonll((uint64_t) atomic_get(&ep->counters[i]));
+		values[i] = htobe64((uint64_t) atomic_get(&ep->counters[i]));
 	*cnt = ACM_MAX_COUNTER;
 }
 
@@ -2082,7 +2082,7 @@  static void acmp_parse_osm_fullv1_lid2guid(FILE *f, uint64_t *lid2guid)
 		if (lid2guid[lid])
 			acm_log(0, "ERROR - duplicate lid %u\n", lid);
 		else
-			lid2guid[lid] = htonll(guid);
+			lid2guid[lid] = htobe64(guid);
 	}
 }
 
diff --git a/ibacm/src/acm.c b/ibacm/src/acm.c
index fd06a2e81fbe76..2cf1ea8b2148c0 100644
--- a/ibacm/src/acm.c
+++ b/ibacm/src/acm.c
@@ -1000,7 +1000,7 @@  static int acm_svr_perf_query(struct acmc_client *client, struct acm_msg *msg)
 	    ((ntohs(msg->hdr.length) >= (ACM_MSG_HDR_LENGTH + ACM_MSG_EP_LENGTH)
 	    && !(msg->resolve_data[0].flags & ACM_EP_FLAG_SOURCE)))) {
 		for (i = 0; i < ACM_MAX_COUNTER; i++)
-			msg->perf_data[i] = htonll((uint64_t) atomic_get(&counter[i]));
+			msg->perf_data[i] = htobe64((uint64_t) atomic_get(&counter[i]));
 
 		msg->hdr.data[0] = ACM_MAX_COUNTER;
 		len = ACM_MSG_HDR_LENGTH + (ACM_MAX_COUNTER * sizeof(uint64_t));
@@ -1426,7 +1426,7 @@  static int acm_nl_parse_path_attr(struct nlattr *attr,
 		sid = (uint64_t *) NLA_DATA(attr);
 		if (NLA_LEN(attr) == sizeof(*sid)) {
 			acm_log(2, "service_id 0x%" PRIx64 "\n", *sid);
-			path->service_id = htonll(*sid);
+			path->service_id = htobe64(*sid);
 		} else {
 			ret = -1;
 		}
@@ -2499,7 +2499,7 @@  static void acm_load_prov_config(void)
 		acm_log(2, "provider %s subnet_prefix 0x%" PRIx64 "\n",
 			prov_name, prefix);
 		/* Convert it into network byte order */
-		prefix = htonll(prefix);
+		prefix = htobe64(prefix);
 
 		list_for_each(&provider_list, prov, entry) {
 			if (!strcasecmp(prov->prov->name, prov_name)) {
diff --git a/providers/cxgb3/cxio_wr.h b/providers/cxgb3/cxio_wr.h
index d56c5298ebaad5..735b64918a15c8 100644
--- a/providers/cxgb3/cxio_wr.h
+++ b/providers/cxgb3/cxio_wr.h
@@ -351,7 +351,7 @@  static inline void build_fw_riwrh(struct fw_riwrh *wqe, enum t3_wr_opcode op,
 	wqe->gen_tid_len = htonl(V_FW_RIWR_GEN(genbit) | V_FW_RIWR_TID(tid) |
 				 V_FW_RIWR_LEN(len));
 	/* 2nd gen bit... */
-        ((union t3_wr *)wqe)->flit[15] = htonll(genbit);
+        ((union t3_wr *)wqe)->flit[15] = htobe64(genbit);
 }
 
 /*
diff --git a/providers/cxgb3/qp.c b/providers/cxgb3/qp.c
index 30dd898f6e874b..b0cf10364e9a6c 100644
--- a/providers/cxgb3/qp.c
+++ b/providers/cxgb3/qp.c
@@ -84,7 +84,7 @@  static inline int iwch_build_rdma_send(union t3_wr *wqe, struct ibv_send_wr *wr,
 			    htonl(wr->sg_list[i].lkey);
 			wqe->send.sgl[i].len =
 			    htonl(wr->sg_list[i].length);
-			wqe->send.sgl[i].to = htonll(wr->sg_list[i].addr);
+			wqe->send.sgl[i].to = htobe64(wr->sg_list[i].addr);
 		}
 		wqe->send.plen = htonl(wqe->send.plen);
 		wqe->send.num_sgle = htonl(wr->num_sge);
@@ -104,7 +104,7 @@  static inline int iwch_build_rdma_write(union t3_wr *wqe,
 	wqe->write.rdmaop = T3_RDMA_WRITE;
 	wqe->write.reserved = 0;
 	wqe->write.stag_sink = htonl(wr->wr.rdma.rkey);
-	wqe->write.to_sink = htonll(wr->wr.rdma.remote_addr);
+	wqe->write.to_sink = htobe64(wr->wr.rdma.remote_addr);
 
 	wqe->write.num_sgle = wr->num_sge;
 
@@ -139,7 +139,7 @@  static inline int iwch_build_rdma_write(union t3_wr *wqe,
 			wqe->write.sgl[i].len =
 			    htonl(wr->sg_list[i].length);
 			wqe->write.sgl[i].to =
-			    htonll(wr->sg_list[i].addr);
+			    htobe64(wr->sg_list[i].addr);
 		}
 		wqe->write.plen = htonl(wqe->write.plen);
 		wqe->write.num_sgle = htonl(wr->num_sge);
@@ -157,10 +157,10 @@  static inline int iwch_build_rdma_read(union t3_wr *wqe, struct ibv_send_wr *wr,
 	wqe->read.reserved = 0;
 	if (wr->num_sge == 1 && wr->sg_list[0].length > 0) {
 		wqe->read.rem_stag = htonl(wr->wr.rdma.rkey);
-		wqe->read.rem_to = htonll(wr->wr.rdma.remote_addr);
+		wqe->read.rem_to = htobe64(wr->wr.rdma.remote_addr);
 		wqe->read.local_stag = htonl(wr->sg_list[0].lkey);
 		wqe->read.local_len = htonl(wr->sg_list[0].length);
-		wqe->read.local_to = htonll(wr->sg_list[0].addr);
+		wqe->read.local_to = htobe64(wr->sg_list[0].addr);
 	} else {
 
 		/* build passable 0B read request */
@@ -295,7 +295,7 @@  static inline int iwch_build_rdma_recv(struct iwch_device *rhp,
 	for (i = 0; i < wr->num_sge; i++) {
 		wqe->recv.sgl[i].stag = htonl(wr->sg_list[i].lkey);
 		wqe->recv.sgl[i].len = htonl(wr->sg_list[i].length);
-		wqe->recv.sgl[i].to = htonll(wr->sg_list[i].addr);
+		wqe->recv.sgl[i].to = htobe64(wr->sg_list[i].addr);
 	}
 	for (; i < T3_MAX_SGE; i++) {
 		wqe->recv.sgl[i].stag = 0;
diff --git a/providers/cxgb4/t4.h b/providers/cxgb4/t4.h
index cad675c3fdb193..59dbc65d4e164b 100644
--- a/providers/cxgb4/t4.h
+++ b/providers/cxgb4/t4.h
@@ -55,7 +55,7 @@ 
 #define __iomem
 #define cpu_to_be16 htons
 #define cpu_to_be32 htonl
-#define cpu_to_be64 htonll
+#define cpu_to_be64 htobe64
 #define be16_to_cpu ntohs
 #define be32_to_cpu ntohl
 #define be64_to_cpu ntohll
diff --git a/providers/mlx4/qp.c b/providers/mlx4/qp.c
index 268fb7dc83dd11..a607326c7c452c 100644
--- a/providers/mlx4/qp.c
+++ b/providers/mlx4/qp.c
@@ -157,7 +157,7 @@  static inline void set_local_inv_seg(struct mlx4_wqe_local_inval_seg *iseg,
 static inline void set_raddr_seg(struct mlx4_wqe_raddr_seg *rseg,
 				 uint64_t remote_addr, uint32_t rkey)
 {
-	rseg->raddr    = htonll(remote_addr);
+	rseg->raddr    = htobe64(remote_addr);
 	rseg->rkey     = htonl(rkey);
 	rseg->reserved = 0;
 }
@@ -165,10 +165,10 @@  static inline void set_raddr_seg(struct mlx4_wqe_raddr_seg *rseg,
 static void set_atomic_seg(struct mlx4_wqe_atomic_seg *aseg, struct ibv_send_wr *wr)
 {
 	if (wr->opcode == IBV_WR_ATOMIC_CMP_AND_SWP) {
-		aseg->swap_add = htonll(wr->wr.atomic.swap);
-		aseg->compare  = htonll(wr->wr.atomic.compare_add);
+		aseg->swap_add = htobe64(wr->wr.atomic.swap);
+		aseg->compare  = htobe64(wr->wr.atomic.compare_add);
 	} else {
-		aseg->swap_add = htonll(wr->wr.atomic.compare_add);
+		aseg->swap_add = htobe64(wr->wr.atomic.compare_add);
 		aseg->compare  = 0;
 	}
 
@@ -188,13 +188,13 @@  static void __set_data_seg(struct mlx4_wqe_data_seg *dseg, struct ibv_sge *sg)
 {
 	dseg->byte_count = htonl(sg->length);
 	dseg->lkey       = htonl(sg->lkey);
-	dseg->addr       = htonll(sg->addr);
+	dseg->addr       = htobe64(sg->addr);
 }
 
 static void set_data_seg(struct mlx4_wqe_data_seg *dseg, struct ibv_sge *sg)
 {
 	dseg->lkey       = htonl(sg->lkey);
-	dseg->addr       = htonll(sg->addr);
+	dseg->addr       = htobe64(sg->addr);
 
 	/*
 	 * Need a barrier here before writing the byte_count field to
diff --git a/providers/mlx4/srq.c b/providers/mlx4/srq.c
index c0e028671828d8..4f90efdf927209 100644
--- a/providers/mlx4/srq.c
+++ b/providers/mlx4/srq.c
@@ -96,7 +96,7 @@  int mlx4_post_srq_recv(struct ibv_srq *ibsrq,
 		for (i = 0; i < wr->num_sge; ++i) {
 			scat[i].byte_count = htonl(wr->sg_list[i].length);
 			scat[i].lkey       = htonl(wr->sg_list[i].lkey);
-			scat[i].addr       = htonll(wr->sg_list[i].addr);
+			scat[i].addr       = htobe64(wr->sg_list[i].addr);
 		}
 
 		if (i < srq->max_gs) {
diff --git a/providers/mlx5/qp.c b/providers/mlx5/qp.c
index e82b1a0bebc3f9..7d034880f4caf6 100644
--- a/providers/mlx5/qp.c
+++ b/providers/mlx5/qp.c
@@ -190,7 +190,7 @@  static int mlx5_wq_overflow(struct mlx5_wq *wq, int nreq, struct mlx5_cq *cq)
 static inline void set_raddr_seg(struct mlx5_wqe_raddr_seg *rseg,
 				 uint64_t remote_addr, uint32_t rkey)
 {
-	rseg->raddr    = htonll(remote_addr);
+	rseg->raddr    = htobe64(remote_addr);
 	rseg->rkey     = htonl(rkey);
 	rseg->reserved = 0;
 }
@@ -201,10 +201,10 @@  static void set_atomic_seg(struct mlx5_wqe_atomic_seg *aseg,
 			   uint64_t compare_add)
 {
 	if (opcode == IBV_WR_ATOMIC_CMP_AND_SWP) {
-		aseg->swap_add = htonll(swap);
-		aseg->compare  = htonll(compare_add);
+		aseg->swap_add = htobe64(swap);
+		aseg->compare  = htobe64(compare_add);
 	} else {
-		aseg->swap_add = htonll(compare_add);
+		aseg->swap_add = htobe64(compare_add);
 	}
 }
 
@@ -221,7 +221,7 @@  static void set_data_ptr_seg(struct mlx5_wqe_data_seg *dseg, struct ibv_sge *sg,
 {
 	dseg->byte_count = htonl(sg->length - offset);
 	dseg->lkey       = htonl(sg->lkey);
-	dseg->addr       = htonll(sg->addr + offset);
+	dseg->addr       = htobe64(sg->addr + offset);
 }
 
 static void set_data_ptr_seg_atomic(struct mlx5_wqe_data_seg *dseg,
@@ -229,7 +229,7 @@  static void set_data_ptr_seg_atomic(struct mlx5_wqe_data_seg *dseg,
 {
 	dseg->byte_count = htonl(MLX5_ATOMIC_SIZE);
 	dseg->lkey       = htonl(sg->lkey);
-	dseg->addr       = htonll(sg->addr);
+	dseg->addr       = htobe64(sg->addr);
 }
 
 /*
@@ -430,7 +430,7 @@  static void set_umr_data_seg(struct mlx5_qp *qp, enum ibv_mw_type type,
 
 	data->klm.byte_count = htonl(bind_info->length);
 	data->klm.mkey = htonl(bind_info->mr->lkey);
-	data->klm.address = htonll(bind_info->addr);
+	data->klm.address = htobe64(bind_info->addr);
 
 	memset(&data->klm + 1, 0, sizeof(data->reserved) -
 	       sizeof(data->klm));
@@ -467,8 +467,8 @@  static void set_umr_mkey_seg(struct mlx5_qp *qp, enum ibv_mw_type type,
 		if (bind_info->mw_access_flags & IBV_ACCESS_ZERO_BASED)
 			mkey->start_addr = 0;
 		else
-			mkey->start_addr = htonll(bind_info->addr);
-		mkey->len = htonll(bind_info->length);
+			mkey->start_addr = htobe64(bind_info->addr);
+		mkey->len = htobe64(bind_info->length);
 	} else {
 		mkey->free = MLX5_WQE_MKEY_CONTEXT_FREE;
 	}
@@ -485,20 +485,20 @@  static inline void set_umr_control_seg(struct mlx5_qp *qp, enum ibv_mw_type type
 
 	ctrl->flags = MLX5_WQE_UMR_CTRL_FLAG_TRNSLATION_OFFSET |
 		MLX5_WQE_UMR_CTRL_FLAG_INLINE;
-	ctrl->mkey_mask = htonll(MLX5_WQE_UMR_CTRL_MKEY_MASK_FREE |
+	ctrl->mkey_mask = htobe64(MLX5_WQE_UMR_CTRL_MKEY_MASK_FREE |
 				     MLX5_WQE_UMR_CTRL_MKEY_MASK_MKEY);
 	ctrl->translation_offset = 0;
 	memset(ctrl->rsvd0, 0, sizeof(ctrl->rsvd0));
 	memset(ctrl->rsvd1, 0, sizeof(ctrl->rsvd1));
 
 	if (type == IBV_MW_TYPE_2)
-		ctrl->mkey_mask |= htonll(MLX5_WQE_UMR_CTRL_MKEY_MASK_QPN);
+		ctrl->mkey_mask |= htobe64(MLX5_WQE_UMR_CTRL_MKEY_MASK_QPN);
 
 	if (bind_info->length) {
 		ctrl->klm_octowords = get_klm_octo(1);
 		if (type == IBV_MW_TYPE_2)
 			ctrl->flags |=  MLX5_WQE_UMR_CTRL_FLAG_CHECK_FREE;
-		ctrl->mkey_mask |= htonll(MLX5_WQE_UMR_CTRL_MKEY_MASK_LEN	|
+		ctrl->mkey_mask |= htobe64(MLX5_WQE_UMR_CTRL_MKEY_MASK_LEN	|
 					      MLX5_WQE_UMR_CTRL_MKEY_MASK_START_ADDR |
 					      MLX5_WQE_UMR_CTRL_MKEY_MASK_ACCESS_LOCAL_WRITE |
 					      MLX5_WQE_UMR_CTRL_MKEY_MASK_ACCESS_REMOTE_READ |
diff --git a/providers/mlx5/srq.c b/providers/mlx5/srq.c
index a06afa3a5931b6..b362ec85724961 100644
--- a/providers/mlx5/srq.c
+++ b/providers/mlx5/srq.c
@@ -120,7 +120,7 @@  int mlx5_post_srq_recv(struct ibv_srq *ibsrq,
 		for (i = 0; i < wr->num_sge; ++i) {
 			scat[i].byte_count = htonl(wr->sg_list[i].length);
 			scat[i].lkey       = htonl(wr->sg_list[i].lkey);
-			scat[i].addr       = htonll(wr->sg_list[i].addr);
+			scat[i].addr       = htobe64(wr->sg_list[i].addr);
 		}
 
 		if (i < srq->max_gs) {
diff --git a/providers/mthca/qp.c b/providers/mthca/qp.c
index 129e6c74ef5382..d221bb19bfa67c 100644
--- a/providers/mthca/qp.c
+++ b/providers/mthca/qp.c
@@ -147,7 +147,7 @@  int mthca_tavor_post_send(struct ibv_qp *ibqp, struct ibv_send_wr *wr,
 			case IBV_WR_ATOMIC_CMP_AND_SWP:
 			case IBV_WR_ATOMIC_FETCH_AND_ADD:
 				((struct mthca_raddr_seg *) wqe)->raddr =
-					htonll(wr->wr.atomic.remote_addr);
+					htobe64(wr->wr.atomic.remote_addr);
 				((struct mthca_raddr_seg *) wqe)->rkey =
 					htonl(wr->wr.atomic.rkey);
 				((struct mthca_raddr_seg *) wqe)->reserved = 0;
@@ -156,12 +156,12 @@  int mthca_tavor_post_send(struct ibv_qp *ibqp, struct ibv_send_wr *wr,
 
 				if (wr->opcode == IBV_WR_ATOMIC_CMP_AND_SWP) {
 					((struct mthca_atomic_seg *) wqe)->swap_add =
-						htonll(wr->wr.atomic.swap);
+						htobe64(wr->wr.atomic.swap);
 					((struct mthca_atomic_seg *) wqe)->compare =
-						htonll(wr->wr.atomic.compare_add);
+						htobe64(wr->wr.atomic.compare_add);
 				} else {
 					((struct mthca_atomic_seg *) wqe)->swap_add =
-						htonll(wr->wr.atomic.compare_add);
+						htobe64(wr->wr.atomic.compare_add);
 					((struct mthca_atomic_seg *) wqe)->compare = 0;
 				}
 
@@ -174,7 +174,7 @@  int mthca_tavor_post_send(struct ibv_qp *ibqp, struct ibv_send_wr *wr,
 			case IBV_WR_RDMA_WRITE_WITH_IMM:
 			case IBV_WR_RDMA_READ:
 				((struct mthca_raddr_seg *) wqe)->raddr =
-					htonll(wr->wr.rdma.remote_addr);
+					htobe64(wr->wr.rdma.remote_addr);
 				((struct mthca_raddr_seg *) wqe)->rkey =
 					htonl(wr->wr.rdma.rkey);
 				((struct mthca_raddr_seg *) wqe)->reserved = 0;
@@ -194,7 +194,7 @@  int mthca_tavor_post_send(struct ibv_qp *ibqp, struct ibv_send_wr *wr,
 			case IBV_WR_RDMA_WRITE:
 			case IBV_WR_RDMA_WRITE_WITH_IMM:
 				((struct mthca_raddr_seg *) wqe)->raddr =
-					htonll(wr->wr.rdma.remote_addr);
+					htobe64(wr->wr.rdma.remote_addr);
 				((struct mthca_raddr_seg *) wqe)->rkey =
 					htonl(wr->wr.rdma.rkey);
 				((struct mthca_raddr_seg *) wqe)->reserved = 0;
@@ -213,7 +213,7 @@  int mthca_tavor_post_send(struct ibv_qp *ibqp, struct ibv_send_wr *wr,
 			((struct mthca_tavor_ud_seg *) wqe)->lkey =
 				htonl(to_mah(wr->wr.ud.ah)->key);
 			((struct mthca_tavor_ud_seg *) wqe)->av_addr =
-				htonll((uintptr_t) to_mah(wr->wr.ud.ah)->av);
+				htobe64((uintptr_t) to_mah(wr->wr.ud.ah)->av);
 			((struct mthca_tavor_ud_seg *) wqe)->dqpn =
 				htonl(wr->wr.ud.remote_qpn);
 			((struct mthca_tavor_ud_seg *) wqe)->qkey =
@@ -265,7 +265,7 @@  int mthca_tavor_post_send(struct ibv_qp *ibqp, struct ibv_send_wr *wr,
 				seg = wqe;
 				seg->byte_count = htonl(wr->sg_list[i].length);
 				seg->lkey = htonl(wr->sg_list[i].lkey);
-				seg->addr = htonll(wr->sg_list[i].addr);
+				seg->addr = htobe64(wr->sg_list[i].addr);
 				wqe += sizeof *seg;
 			}
 
@@ -372,7 +372,7 @@  int mthca_tavor_post_recv(struct ibv_qp *ibqp, struct ibv_recv_wr *wr,
 			((struct mthca_data_seg *) wqe)->lkey =
 				htonl(wr->sg_list[i].lkey);
 			((struct mthca_data_seg *) wqe)->addr =
-				htonll(wr->sg_list[i].addr);
+				htobe64(wr->sg_list[i].addr);
 			wqe += sizeof (struct mthca_data_seg);
 			size += sizeof (struct mthca_data_seg) / 16;
 		}
@@ -508,7 +508,7 @@  int mthca_arbel_post_send(struct ibv_qp *ibqp, struct ibv_send_wr *wr,
 			case IBV_WR_ATOMIC_CMP_AND_SWP:
 			case IBV_WR_ATOMIC_FETCH_AND_ADD:
 				((struct mthca_raddr_seg *) wqe)->raddr =
-					htonll(wr->wr.atomic.remote_addr);
+					htobe64(wr->wr.atomic.remote_addr);
 				((struct mthca_raddr_seg *) wqe)->rkey =
 					htonl(wr->wr.atomic.rkey);
 				((struct mthca_raddr_seg *) wqe)->reserved = 0;
@@ -517,12 +517,12 @@  int mthca_arbel_post_send(struct ibv_qp *ibqp, struct ibv_send_wr *wr,
 
 				if (wr->opcode == IBV_WR_ATOMIC_CMP_AND_SWP) {
 					((struct mthca_atomic_seg *) wqe)->swap_add =
-						htonll(wr->wr.atomic.swap);
+						htobe64(wr->wr.atomic.swap);
 					((struct mthca_atomic_seg *) wqe)->compare =
-						htonll(wr->wr.atomic.compare_add);
+						htobe64(wr->wr.atomic.compare_add);
 				} else {
 					((struct mthca_atomic_seg *) wqe)->swap_add =
-						htonll(wr->wr.atomic.compare_add);
+						htobe64(wr->wr.atomic.compare_add);
 					((struct mthca_atomic_seg *) wqe)->compare = 0;
 				}
 
@@ -535,7 +535,7 @@  int mthca_arbel_post_send(struct ibv_qp *ibqp, struct ibv_send_wr *wr,
 			case IBV_WR_RDMA_WRITE_WITH_IMM:
 			case IBV_WR_RDMA_READ:
 				((struct mthca_raddr_seg *) wqe)->raddr =
-					htonll(wr->wr.rdma.remote_addr);
+					htobe64(wr->wr.rdma.remote_addr);
 				((struct mthca_raddr_seg *) wqe)->rkey =
 					htonl(wr->wr.rdma.rkey);
 				((struct mthca_raddr_seg *) wqe)->reserved = 0;
@@ -555,7 +555,7 @@  int mthca_arbel_post_send(struct ibv_qp *ibqp, struct ibv_send_wr *wr,
 			case IBV_WR_RDMA_WRITE:
 			case IBV_WR_RDMA_WRITE_WITH_IMM:
 				((struct mthca_raddr_seg *) wqe)->raddr =
-					htonll(wr->wr.rdma.remote_addr);
+					htobe64(wr->wr.rdma.remote_addr);
 				((struct mthca_raddr_seg *) wqe)->rkey =
 					htonl(wr->wr.rdma.rkey);
 				((struct mthca_raddr_seg *) wqe)->reserved = 0;
@@ -624,7 +624,7 @@  int mthca_arbel_post_send(struct ibv_qp *ibqp, struct ibv_send_wr *wr,
 				seg = wqe;
 				seg->byte_count = htonl(wr->sg_list[i].length);
 				seg->lkey = htonl(wr->sg_list[i].lkey);
-				seg->addr = htonll(wr->sg_list[i].addr);
+				seg->addr = htobe64(wr->sg_list[i].addr);
 				wqe += sizeof *seg;
 			}
 
@@ -730,7 +730,7 @@  int mthca_arbel_post_recv(struct ibv_qp *ibqp, struct ibv_recv_wr *wr,
 			((struct mthca_data_seg *) wqe)->lkey =
 				htonl(wr->sg_list[i].lkey);
 			((struct mthca_data_seg *) wqe)->addr =
-				htonll(wr->sg_list[i].addr);
+				htobe64(wr->sg_list[i].addr);
 			wqe += sizeof (struct mthca_data_seg);
 		}
 
diff --git a/providers/mthca/srq.c b/providers/mthca/srq.c
index ff6f532733f4ae..66ac924a720c84 100644
--- a/providers/mthca/srq.c
+++ b/providers/mthca/srq.c
@@ -126,7 +126,7 @@  int mthca_tavor_post_srq_recv(struct ibv_srq *ibsrq,
 			((struct mthca_data_seg *) wqe)->lkey =
 				htonl(wr->sg_list[i].lkey);
 			((struct mthca_data_seg *) wqe)->addr =
-				htonll(wr->sg_list[i].addr);
+				htobe64(wr->sg_list[i].addr);
 			wqe += sizeof (struct mthca_data_seg);
 		}
 
@@ -219,7 +219,7 @@  int mthca_arbel_post_srq_recv(struct ibv_srq *ibsrq,
 			((struct mthca_data_seg *) wqe)->lkey =
 				htonl(wr->sg_list[i].lkey);
 			((struct mthca_data_seg *) wqe)->addr =
-				htonll(wr->sg_list[i].addr);
+				htobe64(wr->sg_list[i].addr);
 			wqe += sizeof (struct mthca_data_seg);
 		}
 
diff --git a/providers/qedr/qelr_verbs.c b/providers/qedr/qelr_verbs.c
index 823ed8754c09d7..932115c8ab7f39 100644
--- a/providers/qedr/qelr_verbs.c
+++ b/providers/qedr/qelr_verbs.c
@@ -900,7 +900,7 @@  static inline void qelr_edpm_set_rdma_ext(struct qelr_qp *qp,
 	if (!qp->edpm.is_edpm)
 		return;
 
-	qp->edpm.rdma_ext->remote_va = htonll(remote_addr);
+	qp->edpm.rdma_ext->remote_va = htobe64(remote_addr);
 	qp->edpm.rdma_ext->remote_key = htonl(rkey);
 	qp->edpm.dpm_payload_offset += sizeof(*qp->edpm.rdma_ext);
 	qp->edpm.dpm_payload_size += sizeof(*qp->edpm.rdma_ext);
diff --git a/srp_daemon/srp_daemon.c b/srp_daemon/srp_daemon.c
index e0e550d81ab3c2..ff6118acf611ca 100644
--- a/srp_daemon/srp_daemon.c
+++ b/srp_daemon/srp_daemon.c
@@ -462,9 +462,9 @@  static int add_non_exist_target(struct target_details *target)
 			if (srpd_sys_read_gid(scsi_host_dir, "dgid", dgid_val))
 				continue;
 		}
-		if (htonll(target->subnet_prefix) != *((uint64_t *) dgid_val))
+		if (htobe64(target->subnet_prefix) != *((uint64_t *) dgid_val))
 			continue;
-		if (htonll(target->h_guid) != *((uint64_t *) (dgid_val+8)))
+		if (htobe64(target->h_guid) != *((uint64_t *) (dgid_val+8)))
 			continue;
 
 		/* If there is no local_ib_device in the scsi host dir (old kernel module), assumes it is equal */
@@ -603,7 +603,7 @@  static int send_and_get(int portid, int agent, srp_ib_user_mad_t *out_mad,
 		/* Skip tid 0 because OpenSM ignores it. */
 		if (++tid == 0)
 			++tid;
-		out_dm_mad->tid = htonll(tid);
+		out_dm_mad->tid = htobe64(tid);
 
 		ret = umad_send(portid, agent, out_mad, MAD_BLOCK_SIZE,
 				config->timeout, 0);
@@ -1033,7 +1033,7 @@  int get_node(struct umad_resources *umad_res, uint16_t dlid, uint64_t *guid)
 	init_srp_sa_mad(&out_mad, umad_res->agent, umad_res->sm_lid,
 		        SRP_SA_ATTR_NODE, 0);
 
-	out_sa_mad->comp_mask     = htonll(1); /* LID */
+	out_sa_mad->comp_mask     = htobe64(1); /* LID */
 	node			  = (void *) out_sa_mad->data;
 	node->lid		  = htons(dlid);
 
@@ -1059,7 +1059,7 @@  static int get_port_info(struct umad_resources *umad_res, uint16_t dlid,
 	init_srp_sa_mad(&out_mad, umad_res->agent, umad_res->sm_lid,
 		        SRP_SA_ATTR_PORT_INFO, 0);
 
-	out_sa_mad->comp_mask     = htonll(1); /* LID */
+	out_sa_mad->comp_mask     = htobe64(1); /* LID */
 	port_info                 = (void *) out_sa_mad->data;
 	port_info->endport_lid	  = htons(dlid);
 
@@ -1134,7 +1134,7 @@  static int get_shared_pkeys(struct resources *res,
 			continue;
 
 		/* Mark components: DLID, SLID, PKEY */
-		out_sa_mad->comp_mask = htonll(1 << 4 | 1 << 5 | 1 << 13);
+		out_sa_mad->comp_mask = htobe64(1 << 4 | 1 << 5 | 1 << 13);
 		out_sa_mad->rmpp_version = 1;
 		out_sa_mad->rmpp_type = 1;
 		path_rec = (ib_path_rec_t *)out_sa_mad->data;
@@ -1194,7 +1194,7 @@  static int do_dm_port_list(struct resources *res)
 		        SRP_SA_ATTR_PORT_INFO, SRP_SM_CAP_MASK_MATCH_ATTR_MOD);
 
 	out_sa_mad->method     	   = SRP_SA_METHOD_GET_TABLE;
-	out_sa_mad->comp_mask      = htonll(1 << 7); /* Capability mask */
+	out_sa_mad->comp_mask      = htobe64(1 << 7); /* Capability mask */
 	out_sa_mad->rmpp_version   = 1;
 	out_sa_mad->rmpp_type      = 1;
 	port_info		   = (void *) out_sa_mad->data;
@@ -2330,7 +2330,7 @@  static int get_lid(struct umad_resources *umad_res, ib_gid_t *gid, uint16_t *lid
 	init_srp_sa_mad(&out_mad, umad_res->agent, umad_res->sm_lid,
 		        SRP_SA_ATTR_PATH_REC, 0);
 
-	out_sa_mad->comp_mask = htonll( 4 | 8 | 64 | 512 | 4096 );
+	out_sa_mad->comp_mask = htobe64( 4 | 8 | 64 | 512 | 4096 );
 
 	path_rec->sgid = *gid;
 	path_rec->dgid = *gid;
diff --git a/srp_daemon/srp_handle_traps.c b/srp_daemon/srp_handle_traps.c
index 01cc5445e8352b..6a7a2d1ab479a1 100644
--- a/srp_daemon/srp_handle_traps.c
+++ b/srp_daemon/srp_handle_traps.c
@@ -606,14 +606,14 @@  static int register_to_trap(struct sync_resources *sync_res,
 	    comp_mask |= SRP_INFORMINFO_QPN_COMP;
 	}
 
-	p_sa_mad->comp_mask = htonll(comp_mask);
+	p_sa_mad->comp_mask = htobe64(comp_mask);
 	pr_debug("comp_mask: %llx\n", comp_mask);
 
 	do {
 		pthread_mutex_lock(res->mad_buffer_mutex);
 		res->mad_buffer->base_ver = 0; // flag that the buffer is empty
 		pthread_mutex_unlock(res->mad_buffer_mutex);
-		mad_hdr->trans_id = htonll(trans_id);
+		mad_hdr->trans_id = htobe64(trans_id);
 		trans_id++;
 
 		ret = ibv_post_send(res->qp, &sr, bad_wr);