[go: nahoru, domu]

1/******************************************************************************
2 *
3 * GPL LICENSE SUMMARY
4 *
5 * Copyright(c) 2008 - 2014 Intel Corporation. All rights reserved.
6 *
7 * This program is free software; you can redistribute it and/or modify
8 * it under the terms of version 2 of the GNU General Public License as
9 * published by the Free Software Foundation.
10 *
11 * This program is distributed in the hope that it will be useful, but
12 * WITHOUT ANY WARRANTY; without even the implied warranty of
13 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
14 * General Public License for more details.
15 *
16 * You should have received a copy of the GNU General Public License
17 * along with this program; if not, write to the Free Software
18 * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110,
19 * USA
20 *
21 * The full GNU General Public License is included in this distribution
22 * in the file called COPYING.
23 *
24 * Contact Information:
25 *  Intel Linux Wireless <ilw@linux.intel.com>
26 * Intel Corporation, 5200 N.E. Elam Young Parkway, Hillsboro, OR 97124-6497
27 *
28 *****************************************************************************/
29#include <linux/etherdevice.h>
30#include <linux/kernel.h>
31#include <linux/module.h>
32#include <linux/sched.h>
33#include <net/mac80211.h>
34
35#include "iwl-io.h"
36#include "iwl-agn-hw.h"
37#include "iwl-trans.h"
38#include "iwl-modparams.h"
39
40#include "dev.h"
41#include "agn.h"
42
43int iwlagn_hw_valid_rtc_data_addr(u32 addr)
44{
45	return (addr >= IWLAGN_RTC_DATA_LOWER_BOUND) &&
46		(addr < IWLAGN_RTC_DATA_UPPER_BOUND);
47}
48
49int iwlagn_send_tx_power(struct iwl_priv *priv)
50{
51	struct iwlagn_tx_power_dbm_cmd tx_power_cmd;
52	u8 tx_ant_cfg_cmd;
53
54	if (WARN_ONCE(test_bit(STATUS_SCAN_HW, &priv->status),
55		      "TX Power requested while scanning!\n"))
56		return -EAGAIN;
57
58	/* half dBm need to multiply */
59	tx_power_cmd.global_lmt = (s8)(2 * priv->tx_power_user_lmt);
60
61	if (tx_power_cmd.global_lmt > priv->nvm_data->max_tx_pwr_half_dbm) {
62		/*
63		 * For the newer devices which using enhanced/extend tx power
64		 * table in EEPROM, the format is in half dBm. driver need to
65		 * convert to dBm format before report to mac80211.
66		 * By doing so, there is a possibility of 1/2 dBm resolution
67		 * lost. driver will perform "round-up" operation before
68		 * reporting, but it will cause 1/2 dBm tx power over the
69		 * regulatory limit. Perform the checking here, if the
70		 * "tx_power_user_lmt" is higher than EEPROM value (in
71		 * half-dBm format), lower the tx power based on EEPROM
72		 */
73		tx_power_cmd.global_lmt =
74			priv->nvm_data->max_tx_pwr_half_dbm;
75	}
76	tx_power_cmd.flags = IWLAGN_TX_POWER_NO_CLOSED;
77	tx_power_cmd.srv_chan_lmt = IWLAGN_TX_POWER_AUTO;
78
79	if (IWL_UCODE_API(priv->fw->ucode_ver) == 1)
80		tx_ant_cfg_cmd = REPLY_TX_POWER_DBM_CMD_V1;
81	else
82		tx_ant_cfg_cmd = REPLY_TX_POWER_DBM_CMD;
83
84	return iwl_dvm_send_cmd_pdu(priv, tx_ant_cfg_cmd, 0,
85			sizeof(tx_power_cmd), &tx_power_cmd);
86}
87
88void iwlagn_temperature(struct iwl_priv *priv)
89{
90	lockdep_assert_held(&priv->statistics.lock);
91
92	/* store temperature from correct statistics (in Celsius) */
93	priv->temperature = le32_to_cpu(priv->statistics.common.temperature);
94	iwl_tt_handler(priv);
95}
96
97int iwlagn_hwrate_to_mac80211_idx(u32 rate_n_flags, enum ieee80211_band band)
98{
99	int idx = 0;
100	int band_offset = 0;
101
102	/* HT rate format: mac80211 wants an MCS number, which is just LSB */
103	if (rate_n_flags & RATE_MCS_HT_MSK) {
104		idx = (rate_n_flags & 0xff);
105		return idx;
106	/* Legacy rate format, search for match in table */
107	} else {
108		if (band == IEEE80211_BAND_5GHZ)
109			band_offset = IWL_FIRST_OFDM_RATE;
110		for (idx = band_offset; idx < IWL_RATE_COUNT_LEGACY; idx++)
111			if (iwl_rates[idx].plcp == (rate_n_flags & 0xFF))
112				return idx - band_offset;
113	}
114
115	return -1;
116}
117
118int iwlagn_manage_ibss_station(struct iwl_priv *priv,
119			       struct ieee80211_vif *vif, bool add)
120{
121	struct iwl_vif_priv *vif_priv = (void *)vif->drv_priv;
122
123	if (add)
124		return iwlagn_add_bssid_station(priv, vif_priv->ctx,
125						vif->bss_conf.bssid,
126						&vif_priv->ibss_bssid_sta_id);
127	return iwl_remove_station(priv, vif_priv->ibss_bssid_sta_id,
128				  vif->bss_conf.bssid);
129}
130
131/**
132 * iwlagn_txfifo_flush: send REPLY_TXFIFO_FLUSH command to uCode
133 *
134 * pre-requirements:
135 *  1. acquire mutex before calling
136 *  2. make sure rf is on and not in exit state
137 */
138int iwlagn_txfifo_flush(struct iwl_priv *priv, u32 scd_q_msk)
139{
140	struct iwl_txfifo_flush_cmd flush_cmd;
141	struct iwl_host_cmd cmd = {
142		.id = REPLY_TXFIFO_FLUSH,
143		.len = { sizeof(struct iwl_txfifo_flush_cmd), },
144		.data = { &flush_cmd, },
145	};
146
147	memset(&flush_cmd, 0, sizeof(flush_cmd));
148
149	flush_cmd.queue_control = IWL_SCD_VO_MSK | IWL_SCD_VI_MSK |
150				  IWL_SCD_BE_MSK | IWL_SCD_BK_MSK |
151				  IWL_SCD_MGMT_MSK;
152	if ((priv->valid_contexts != BIT(IWL_RXON_CTX_BSS)))
153		flush_cmd.queue_control |= IWL_PAN_SCD_VO_MSK |
154					   IWL_PAN_SCD_VI_MSK |
155					   IWL_PAN_SCD_BE_MSK |
156					   IWL_PAN_SCD_BK_MSK |
157					   IWL_PAN_SCD_MGMT_MSK |
158					   IWL_PAN_SCD_MULTICAST_MSK;
159
160	if (priv->nvm_data->sku_cap_11n_enable)
161		flush_cmd.queue_control |= IWL_AGG_TX_QUEUE_MSK;
162
163	if (scd_q_msk)
164		flush_cmd.queue_control = cpu_to_le32(scd_q_msk);
165
166	IWL_DEBUG_INFO(priv, "queue control: 0x%x\n",
167		       flush_cmd.queue_control);
168	flush_cmd.flush_control = cpu_to_le16(IWL_DROP_ALL);
169
170	return iwl_dvm_send_cmd(priv, &cmd);
171}
172
173void iwlagn_dev_txfifo_flush(struct iwl_priv *priv)
174{
175	mutex_lock(&priv->mutex);
176	ieee80211_stop_queues(priv->hw);
177	if (iwlagn_txfifo_flush(priv, 0)) {
178		IWL_ERR(priv, "flush request fail\n");
179		goto done;
180	}
181	IWL_DEBUG_INFO(priv, "wait transmit/flush all frames\n");
182	iwl_trans_wait_tx_queue_empty(priv->trans, 0xffffffff);
183done:
184	ieee80211_wake_queues(priv->hw);
185	mutex_unlock(&priv->mutex);
186}
187
188/*
189 * BT coex
190 */
191/* Notmal TDM */
192static const __le32 iwlagn_def_3w_lookup[IWLAGN_BT_DECISION_LUT_SIZE] = {
193	cpu_to_le32(0xaaaaaaaa),
194	cpu_to_le32(0xaaaaaaaa),
195	cpu_to_le32(0xaeaaaaaa),
196	cpu_to_le32(0xaaaaaaaa),
197	cpu_to_le32(0xcc00ff28),
198	cpu_to_le32(0x0000aaaa),
199	cpu_to_le32(0xcc00aaaa),
200	cpu_to_le32(0x0000aaaa),
201	cpu_to_le32(0xc0004000),
202	cpu_to_le32(0x00004000),
203	cpu_to_le32(0xf0005000),
204	cpu_to_le32(0xf0005000),
205};
206
207
208/* Loose Coex */
209static const __le32 iwlagn_loose_lookup[IWLAGN_BT_DECISION_LUT_SIZE] = {
210	cpu_to_le32(0xaaaaaaaa),
211	cpu_to_le32(0xaaaaaaaa),
212	cpu_to_le32(0xaeaaaaaa),
213	cpu_to_le32(0xaaaaaaaa),
214	cpu_to_le32(0xcc00ff28),
215	cpu_to_le32(0x0000aaaa),
216	cpu_to_le32(0xcc00aaaa),
217	cpu_to_le32(0x0000aaaa),
218	cpu_to_le32(0x00000000),
219	cpu_to_le32(0x00000000),
220	cpu_to_le32(0xf0005000),
221	cpu_to_le32(0xf0005000),
222};
223
224/* Full concurrency */
225static const __le32 iwlagn_concurrent_lookup[IWLAGN_BT_DECISION_LUT_SIZE] = {
226	cpu_to_le32(0xaaaaaaaa),
227	cpu_to_le32(0xaaaaaaaa),
228	cpu_to_le32(0xaaaaaaaa),
229	cpu_to_le32(0xaaaaaaaa),
230	cpu_to_le32(0xaaaaaaaa),
231	cpu_to_le32(0xaaaaaaaa),
232	cpu_to_le32(0xaaaaaaaa),
233	cpu_to_le32(0xaaaaaaaa),
234	cpu_to_le32(0x00000000),
235	cpu_to_le32(0x00000000),
236	cpu_to_le32(0x00000000),
237	cpu_to_le32(0x00000000),
238};
239
240void iwlagn_send_advance_bt_config(struct iwl_priv *priv)
241{
242	struct iwl_basic_bt_cmd basic = {
243		.max_kill = IWLAGN_BT_MAX_KILL_DEFAULT,
244		.bt3_timer_t7_value = IWLAGN_BT3_T7_DEFAULT,
245		.bt3_prio_sample_time = IWLAGN_BT3_PRIO_SAMPLE_DEFAULT,
246		.bt3_timer_t2_value = IWLAGN_BT3_T2_DEFAULT,
247	};
248	struct iwl_bt_cmd_v1 bt_cmd_v1;
249	struct iwl_bt_cmd_v2 bt_cmd_v2;
250	int ret;
251
252	BUILD_BUG_ON(sizeof(iwlagn_def_3w_lookup) !=
253			sizeof(basic.bt3_lookup_table));
254
255	if (priv->lib->bt_params) {
256		/*
257		 * newer generation of devices (2000 series and newer)
258		 * use the version 2 of the bt command
259		 * we need to make sure sending the host command
260		 * with correct data structure to avoid uCode assert
261		 */
262		if (priv->lib->bt_params->bt_session_2) {
263			bt_cmd_v2.prio_boost = cpu_to_le32(
264				priv->lib->bt_params->bt_prio_boost);
265			bt_cmd_v2.tx_prio_boost = 0;
266			bt_cmd_v2.rx_prio_boost = 0;
267		} else {
268			/* older version only has 8 bits */
269			WARN_ON(priv->lib->bt_params->bt_prio_boost & ~0xFF);
270			bt_cmd_v1.prio_boost =
271				priv->lib->bt_params->bt_prio_boost;
272			bt_cmd_v1.tx_prio_boost = 0;
273			bt_cmd_v1.rx_prio_boost = 0;
274		}
275	} else {
276		IWL_ERR(priv, "failed to construct BT Coex Config\n");
277		return;
278	}
279
280	/*
281	 * Possible situations when BT needs to take over for receive,
282	 * at the same time where STA needs to response to AP's frame(s),
283	 * reduce the tx power of the required response frames, by that,
284	 * allow the concurrent BT receive & WiFi transmit
285	 * (BT - ANT A, WiFi -ANT B), without interference to one another
286	 *
287	 * Reduced tx power apply to control frames only (ACK/Back/CTS)
288	 * when indicated by the BT config command
289	 */
290	basic.kill_ack_mask = priv->kill_ack_mask;
291	basic.kill_cts_mask = priv->kill_cts_mask;
292	if (priv->reduced_txpower)
293		basic.reduce_txpower = IWLAGN_BT_REDUCED_TX_PWR;
294	basic.valid = priv->bt_valid;
295
296	/*
297	 * Configure BT coex mode to "no coexistence" when the
298	 * user disabled BT coexistence, we have no interface
299	 * (might be in monitor mode), or the interface is in
300	 * IBSS mode (no proper uCode support for coex then).
301	 */
302	if (!iwlwifi_mod_params.bt_coex_active ||
303	    priv->iw_mode == NL80211_IFTYPE_ADHOC) {
304		basic.flags = IWLAGN_BT_FLAG_COEX_MODE_DISABLED;
305	} else {
306		basic.flags = IWLAGN_BT_FLAG_COEX_MODE_3W <<
307					IWLAGN_BT_FLAG_COEX_MODE_SHIFT;
308
309		if (!priv->bt_enable_pspoll)
310			basic.flags |= IWLAGN_BT_FLAG_SYNC_2_BT_DISABLE;
311		else
312			basic.flags &= ~IWLAGN_BT_FLAG_SYNC_2_BT_DISABLE;
313
314		if (priv->bt_ch_announce)
315			basic.flags |= IWLAGN_BT_FLAG_CHANNEL_INHIBITION;
316		IWL_DEBUG_COEX(priv, "BT coex flag: 0X%x\n", basic.flags);
317	}
318	priv->bt_enable_flag = basic.flags;
319	if (priv->bt_full_concurrent)
320		memcpy(basic.bt3_lookup_table, iwlagn_concurrent_lookup,
321			sizeof(iwlagn_concurrent_lookup));
322	else
323		memcpy(basic.bt3_lookup_table, iwlagn_def_3w_lookup,
324			sizeof(iwlagn_def_3w_lookup));
325
326	IWL_DEBUG_COEX(priv, "BT coex %s in %s mode\n",
327		       basic.flags ? "active" : "disabled",
328		       priv->bt_full_concurrent ?
329		       "full concurrency" : "3-wire");
330
331	if (priv->lib->bt_params->bt_session_2) {
332		memcpy(&bt_cmd_v2.basic, &basic,
333			sizeof(basic));
334		ret = iwl_dvm_send_cmd_pdu(priv, REPLY_BT_CONFIG,
335			0, sizeof(bt_cmd_v2), &bt_cmd_v2);
336	} else {
337		memcpy(&bt_cmd_v1.basic, &basic,
338			sizeof(basic));
339		ret = iwl_dvm_send_cmd_pdu(priv, REPLY_BT_CONFIG,
340			0, sizeof(bt_cmd_v1), &bt_cmd_v1);
341	}
342	if (ret)
343		IWL_ERR(priv, "failed to send BT Coex Config\n");
344
345}
346
347void iwlagn_bt_adjust_rssi_monitor(struct iwl_priv *priv, bool rssi_ena)
348{
349	struct iwl_rxon_context *ctx, *found_ctx = NULL;
350	bool found_ap = false;
351
352	lockdep_assert_held(&priv->mutex);
353
354	/* Check whether AP or GO mode is active. */
355	if (rssi_ena) {
356		for_each_context(priv, ctx) {
357			if (ctx->vif && ctx->vif->type == NL80211_IFTYPE_AP &&
358			    iwl_is_associated_ctx(ctx)) {
359				found_ap = true;
360				break;
361			}
362		}
363	}
364
365	/*
366	 * If disable was received or If GO/AP mode, disable RSSI
367	 * measurements.
368	 */
369	if (!rssi_ena || found_ap) {
370		if (priv->cur_rssi_ctx) {
371			ctx = priv->cur_rssi_ctx;
372			ieee80211_disable_rssi_reports(ctx->vif);
373			priv->cur_rssi_ctx = NULL;
374		}
375		return;
376	}
377
378	/*
379	 * If rssi measurements need to be enabled, consider all cases now.
380	 * Figure out how many contexts are active.
381	 */
382	for_each_context(priv, ctx) {
383		if (ctx->vif && ctx->vif->type == NL80211_IFTYPE_STATION &&
384		    iwl_is_associated_ctx(ctx)) {
385			found_ctx = ctx;
386			break;
387		}
388	}
389
390	/*
391	 * rssi monitor already enabled for the correct interface...nothing
392	 * to do.
393	 */
394	if (found_ctx == priv->cur_rssi_ctx)
395		return;
396
397	/*
398	 * Figure out if rssi monitor is currently enabled, and needs
399	 * to be changed. If rssi monitor is already enabled, disable
400	 * it first else just enable rssi measurements on the
401	 * interface found above.
402	 */
403	if (priv->cur_rssi_ctx) {
404		ctx = priv->cur_rssi_ctx;
405		if (ctx->vif)
406			ieee80211_disable_rssi_reports(ctx->vif);
407	}
408
409	priv->cur_rssi_ctx = found_ctx;
410
411	if (!found_ctx)
412		return;
413
414	ieee80211_enable_rssi_reports(found_ctx->vif,
415			IWLAGN_BT_PSP_MIN_RSSI_THRESHOLD,
416			IWLAGN_BT_PSP_MAX_RSSI_THRESHOLD);
417}
418
419static bool iwlagn_bt_traffic_is_sco(struct iwl_bt_uart_msg *uart_msg)
420{
421	return BT_UART_MSG_FRAME3SCOESCO_MSK & uart_msg->frame3 >>
422			BT_UART_MSG_FRAME3SCOESCO_POS;
423}
424
425static void iwlagn_bt_traffic_change_work(struct work_struct *work)
426{
427	struct iwl_priv *priv =
428		container_of(work, struct iwl_priv, bt_traffic_change_work);
429	struct iwl_rxon_context *ctx;
430	int smps_request = -1;
431
432	if (priv->bt_enable_flag == IWLAGN_BT_FLAG_COEX_MODE_DISABLED) {
433		/* bt coex disabled */
434		return;
435	}
436
437	/*
438	 * Note: bt_traffic_load can be overridden by scan complete and
439	 * coex profile notifications. Ignore that since only bad consequence
440	 * can be not matching debug print with actual state.
441	 */
442	IWL_DEBUG_COEX(priv, "BT traffic load changes: %d\n",
443		       priv->bt_traffic_load);
444
445	switch (priv->bt_traffic_load) {
446	case IWL_BT_COEX_TRAFFIC_LOAD_NONE:
447		if (priv->bt_status)
448			smps_request = IEEE80211_SMPS_DYNAMIC;
449		else
450			smps_request = IEEE80211_SMPS_AUTOMATIC;
451		break;
452	case IWL_BT_COEX_TRAFFIC_LOAD_LOW:
453		smps_request = IEEE80211_SMPS_DYNAMIC;
454		break;
455	case IWL_BT_COEX_TRAFFIC_LOAD_HIGH:
456	case IWL_BT_COEX_TRAFFIC_LOAD_CONTINUOUS:
457		smps_request = IEEE80211_SMPS_STATIC;
458		break;
459	default:
460		IWL_ERR(priv, "Invalid BT traffic load: %d\n",
461			priv->bt_traffic_load);
462		break;
463	}
464
465	mutex_lock(&priv->mutex);
466
467	/*
468	 * We can not send command to firmware while scanning. When the scan
469	 * complete we will schedule this work again. We do check with mutex
470	 * locked to prevent new scan request to arrive. We do not check
471	 * STATUS_SCANNING to avoid race when queue_work two times from
472	 * different notifications, but quit and not perform any work at all.
473	 */
474	if (test_bit(STATUS_SCAN_HW, &priv->status))
475		goto out;
476
477	iwl_update_chain_flags(priv);
478
479	if (smps_request != -1) {
480		priv->current_ht_config.smps = smps_request;
481		for_each_context(priv, ctx) {
482			if (ctx->vif && ctx->vif->type == NL80211_IFTYPE_STATION)
483				ieee80211_request_smps(ctx->vif, smps_request);
484		}
485	}
486
487	/*
488	 * Dynamic PS poll related functionality. Adjust RSSI measurements if
489	 * necessary.
490	 */
491	iwlagn_bt_coex_rssi_monitor(priv);
492out:
493	mutex_unlock(&priv->mutex);
494}
495
496/*
497 * If BT sco traffic, and RSSI monitor is enabled, move measurements to the
498 * correct interface or disable it if this is the last interface to be
499 * removed.
500 */
501void iwlagn_bt_coex_rssi_monitor(struct iwl_priv *priv)
502{
503	if (priv->bt_is_sco &&
504	    priv->bt_traffic_load == IWL_BT_COEX_TRAFFIC_LOAD_CONTINUOUS)
505		iwlagn_bt_adjust_rssi_monitor(priv, true);
506	else
507		iwlagn_bt_adjust_rssi_monitor(priv, false);
508}
509
510static void iwlagn_print_uartmsg(struct iwl_priv *priv,
511				struct iwl_bt_uart_msg *uart_msg)
512{
513	IWL_DEBUG_COEX(priv, "Message Type = 0x%X, SSN = 0x%X, "
514			"Update Req = 0x%X\n",
515		(BT_UART_MSG_FRAME1MSGTYPE_MSK & uart_msg->frame1) >>
516			BT_UART_MSG_FRAME1MSGTYPE_POS,
517		(BT_UART_MSG_FRAME1SSN_MSK & uart_msg->frame1) >>
518			BT_UART_MSG_FRAME1SSN_POS,
519		(BT_UART_MSG_FRAME1UPDATEREQ_MSK & uart_msg->frame1) >>
520			BT_UART_MSG_FRAME1UPDATEREQ_POS);
521
522	IWL_DEBUG_COEX(priv, "Open connections = 0x%X, Traffic load = 0x%X, "
523			"Chl_SeqN = 0x%X, In band = 0x%X\n",
524		(BT_UART_MSG_FRAME2OPENCONNECTIONS_MSK & uart_msg->frame2) >>
525			BT_UART_MSG_FRAME2OPENCONNECTIONS_POS,
526		(BT_UART_MSG_FRAME2TRAFFICLOAD_MSK & uart_msg->frame2) >>
527			BT_UART_MSG_FRAME2TRAFFICLOAD_POS,
528		(BT_UART_MSG_FRAME2CHLSEQN_MSK & uart_msg->frame2) >>
529			BT_UART_MSG_FRAME2CHLSEQN_POS,
530		(BT_UART_MSG_FRAME2INBAND_MSK & uart_msg->frame2) >>
531			BT_UART_MSG_FRAME2INBAND_POS);
532
533	IWL_DEBUG_COEX(priv, "SCO/eSCO = 0x%X, Sniff = 0x%X, A2DP = 0x%X, "
534			"ACL = 0x%X, Master = 0x%X, OBEX = 0x%X\n",
535		(BT_UART_MSG_FRAME3SCOESCO_MSK & uart_msg->frame3) >>
536			BT_UART_MSG_FRAME3SCOESCO_POS,
537		(BT_UART_MSG_FRAME3SNIFF_MSK & uart_msg->frame3) >>
538			BT_UART_MSG_FRAME3SNIFF_POS,
539		(BT_UART_MSG_FRAME3A2DP_MSK & uart_msg->frame3) >>
540			BT_UART_MSG_FRAME3A2DP_POS,
541		(BT_UART_MSG_FRAME3ACL_MSK & uart_msg->frame3) >>
542			BT_UART_MSG_FRAME3ACL_POS,
543		(BT_UART_MSG_FRAME3MASTER_MSK & uart_msg->frame3) >>
544			BT_UART_MSG_FRAME3MASTER_POS,
545		(BT_UART_MSG_FRAME3OBEX_MSK & uart_msg->frame3) >>
546			BT_UART_MSG_FRAME3OBEX_POS);
547
548	IWL_DEBUG_COEX(priv, "Idle duration = 0x%X\n",
549		(BT_UART_MSG_FRAME4IDLEDURATION_MSK & uart_msg->frame4) >>
550			BT_UART_MSG_FRAME4IDLEDURATION_POS);
551
552	IWL_DEBUG_COEX(priv, "Tx Activity = 0x%X, Rx Activity = 0x%X, "
553			"eSCO Retransmissions = 0x%X\n",
554		(BT_UART_MSG_FRAME5TXACTIVITY_MSK & uart_msg->frame5) >>
555			BT_UART_MSG_FRAME5TXACTIVITY_POS,
556		(BT_UART_MSG_FRAME5RXACTIVITY_MSK & uart_msg->frame5) >>
557			BT_UART_MSG_FRAME5RXACTIVITY_POS,
558		(BT_UART_MSG_FRAME5ESCORETRANSMIT_MSK & uart_msg->frame5) >>
559			BT_UART_MSG_FRAME5ESCORETRANSMIT_POS);
560
561	IWL_DEBUG_COEX(priv, "Sniff Interval = 0x%X, Discoverable = 0x%X\n",
562		(BT_UART_MSG_FRAME6SNIFFINTERVAL_MSK & uart_msg->frame6) >>
563			BT_UART_MSG_FRAME6SNIFFINTERVAL_POS,
564		(BT_UART_MSG_FRAME6DISCOVERABLE_MSK & uart_msg->frame6) >>
565			BT_UART_MSG_FRAME6DISCOVERABLE_POS);
566
567	IWL_DEBUG_COEX(priv, "Sniff Activity = 0x%X, Page = "
568			"0x%X, Inquiry = 0x%X, Connectable = 0x%X\n",
569		(BT_UART_MSG_FRAME7SNIFFACTIVITY_MSK & uart_msg->frame7) >>
570			BT_UART_MSG_FRAME7SNIFFACTIVITY_POS,
571		(BT_UART_MSG_FRAME7PAGE_MSK & uart_msg->frame7) >>
572			BT_UART_MSG_FRAME7PAGE_POS,
573		(BT_UART_MSG_FRAME7INQUIRY_MSK & uart_msg->frame7) >>
574			BT_UART_MSG_FRAME7INQUIRY_POS,
575		(BT_UART_MSG_FRAME7CONNECTABLE_MSK & uart_msg->frame7) >>
576			BT_UART_MSG_FRAME7CONNECTABLE_POS);
577}
578
579static bool iwlagn_set_kill_msk(struct iwl_priv *priv,
580				struct iwl_bt_uart_msg *uart_msg)
581{
582	bool need_update = false;
583	u8 kill_msk = IWL_BT_KILL_REDUCE;
584	static const __le32 bt_kill_ack_msg[3] = {
585		IWLAGN_BT_KILL_ACK_MASK_DEFAULT,
586		IWLAGN_BT_KILL_ACK_CTS_MASK_SCO,
587		IWLAGN_BT_KILL_ACK_CTS_MASK_REDUCE};
588	static const __le32 bt_kill_cts_msg[3] = {
589		IWLAGN_BT_KILL_CTS_MASK_DEFAULT,
590		IWLAGN_BT_KILL_ACK_CTS_MASK_SCO,
591		IWLAGN_BT_KILL_ACK_CTS_MASK_REDUCE};
592
593	if (!priv->reduced_txpower)
594		kill_msk = (BT_UART_MSG_FRAME3SCOESCO_MSK & uart_msg->frame3)
595			? IWL_BT_KILL_OVERRIDE : IWL_BT_KILL_DEFAULT;
596	if (priv->kill_ack_mask != bt_kill_ack_msg[kill_msk] ||
597	    priv->kill_cts_mask != bt_kill_cts_msg[kill_msk]) {
598		priv->bt_valid |= IWLAGN_BT_VALID_KILL_ACK_MASK;
599		priv->kill_ack_mask = bt_kill_ack_msg[kill_msk];
600		priv->bt_valid |= IWLAGN_BT_VALID_KILL_CTS_MASK;
601		priv->kill_cts_mask = bt_kill_cts_msg[kill_msk];
602		need_update = true;
603	}
604	return need_update;
605}
606
607/*
608 * Upon RSSI changes, sends a bt config command with following changes
609 *  1. enable/disable "reduced control frames tx power
610 *  2. update the "kill)ack_mask" and "kill_cts_mask"
611 *
612 * If "reduced tx power" is enabled, uCode shall
613 *  1. ACK/Back/CTS rate shall reduced to 6Mbps
614 *  2. not use duplciate 20/40MHz mode
615 */
616static bool iwlagn_fill_txpower_mode(struct iwl_priv *priv,
617				struct iwl_bt_uart_msg *uart_msg)
618{
619	bool need_update = false;
620	struct iwl_rxon_context *ctx = &priv->contexts[IWL_RXON_CTX_BSS];
621	int ave_rssi;
622
623	if (!ctx->vif || (ctx->vif->type != NL80211_IFTYPE_STATION)) {
624		IWL_DEBUG_INFO(priv, "BSS ctx not active or not in sta mode\n");
625		return false;
626	}
627
628	ave_rssi = ieee80211_ave_rssi(ctx->vif);
629	if (!ave_rssi) {
630		/* no rssi data, no changes to reduce tx power */
631		IWL_DEBUG_COEX(priv, "no rssi data available\n");
632		return need_update;
633	}
634	if (!priv->reduced_txpower &&
635	    !iwl_is_associated(priv, IWL_RXON_CTX_PAN) &&
636	    (ave_rssi > BT_ENABLE_REDUCED_TXPOWER_THRESHOLD) &&
637	    (uart_msg->frame3 & (BT_UART_MSG_FRAME3ACL_MSK |
638	    BT_UART_MSG_FRAME3OBEX_MSK)) &&
639	    !(uart_msg->frame3 & (BT_UART_MSG_FRAME3SCOESCO_MSK |
640	    BT_UART_MSG_FRAME3SNIFF_MSK | BT_UART_MSG_FRAME3A2DP_MSK))) {
641		/* enabling reduced tx power */
642		priv->reduced_txpower = true;
643		priv->bt_valid |= IWLAGN_BT_VALID_REDUCED_TX_PWR;
644		need_update = true;
645	} else if (priv->reduced_txpower &&
646		   (iwl_is_associated(priv, IWL_RXON_CTX_PAN) ||
647		   (ave_rssi < BT_DISABLE_REDUCED_TXPOWER_THRESHOLD) ||
648		   (uart_msg->frame3 & (BT_UART_MSG_FRAME3SCOESCO_MSK |
649		   BT_UART_MSG_FRAME3SNIFF_MSK | BT_UART_MSG_FRAME3A2DP_MSK)) ||
650		   !(uart_msg->frame3 & (BT_UART_MSG_FRAME3ACL_MSK |
651		   BT_UART_MSG_FRAME3OBEX_MSK)))) {
652		/* disable reduced tx power */
653		priv->reduced_txpower = false;
654		priv->bt_valid |= IWLAGN_BT_VALID_REDUCED_TX_PWR;
655		need_update = true;
656	}
657
658	return need_update;
659}
660
661int iwlagn_bt_coex_profile_notif(struct iwl_priv *priv,
662				  struct iwl_rx_cmd_buffer *rxb,
663				  struct iwl_device_cmd *cmd)
664{
665	struct iwl_rx_packet *pkt = rxb_addr(rxb);
666	struct iwl_bt_coex_profile_notif *coex = (void *)pkt->data;
667	struct iwl_bt_uart_msg *uart_msg = &coex->last_bt_uart_msg;
668
669	if (priv->bt_enable_flag == IWLAGN_BT_FLAG_COEX_MODE_DISABLED) {
670		/* bt coex disabled */
671		return 0;
672	}
673
674	IWL_DEBUG_COEX(priv, "BT Coex notification:\n");
675	IWL_DEBUG_COEX(priv, "    status: %d\n", coex->bt_status);
676	IWL_DEBUG_COEX(priv, "    traffic load: %d\n", coex->bt_traffic_load);
677	IWL_DEBUG_COEX(priv, "    CI compliance: %d\n",
678			coex->bt_ci_compliance);
679	iwlagn_print_uartmsg(priv, uart_msg);
680
681	priv->last_bt_traffic_load = priv->bt_traffic_load;
682	priv->bt_is_sco = iwlagn_bt_traffic_is_sco(uart_msg);
683
684	if (priv->iw_mode != NL80211_IFTYPE_ADHOC) {
685		if (priv->bt_status != coex->bt_status ||
686		    priv->last_bt_traffic_load != coex->bt_traffic_load) {
687			if (coex->bt_status) {
688				/* BT on */
689				if (!priv->bt_ch_announce)
690					priv->bt_traffic_load =
691						IWL_BT_COEX_TRAFFIC_LOAD_HIGH;
692				else
693					priv->bt_traffic_load =
694						coex->bt_traffic_load;
695			} else {
696				/* BT off */
697				priv->bt_traffic_load =
698					IWL_BT_COEX_TRAFFIC_LOAD_NONE;
699			}
700			priv->bt_status = coex->bt_status;
701			queue_work(priv->workqueue,
702				   &priv->bt_traffic_change_work);
703		}
704	}
705
706	/* schedule to send runtime bt_config */
707	/* check reduce power before change ack/cts kill mask */
708	if (iwlagn_fill_txpower_mode(priv, uart_msg) ||
709	    iwlagn_set_kill_msk(priv, uart_msg))
710		queue_work(priv->workqueue, &priv->bt_runtime_config);
711
712
713	/* FIXME: based on notification, adjust the prio_boost */
714
715	priv->bt_ci_compliance = coex->bt_ci_compliance;
716	return 0;
717}
718
719void iwlagn_bt_rx_handler_setup(struct iwl_priv *priv)
720{
721	priv->rx_handlers[REPLY_BT_COEX_PROFILE_NOTIF] =
722		iwlagn_bt_coex_profile_notif;
723}
724
725void iwlagn_bt_setup_deferred_work(struct iwl_priv *priv)
726{
727	INIT_WORK(&priv->bt_traffic_change_work,
728		  iwlagn_bt_traffic_change_work);
729}
730
731void iwlagn_bt_cancel_deferred_work(struct iwl_priv *priv)
732{
733	cancel_work_sync(&priv->bt_traffic_change_work);
734}
735
736static bool is_single_rx_stream(struct iwl_priv *priv)
737{
738	return priv->current_ht_config.smps == IEEE80211_SMPS_STATIC ||
739	       priv->current_ht_config.single_chain_sufficient;
740}
741
742#define IWL_NUM_RX_CHAINS_MULTIPLE	3
743#define IWL_NUM_RX_CHAINS_SINGLE	2
744#define IWL_NUM_IDLE_CHAINS_DUAL	2
745#define IWL_NUM_IDLE_CHAINS_SINGLE	1
746
747/*
748 * Determine how many receiver/antenna chains to use.
749 *
750 * More provides better reception via diversity.  Fewer saves power
751 * at the expense of throughput, but only when not in powersave to
752 * start with.
753 *
754 * MIMO (dual stream) requires at least 2, but works better with 3.
755 * This does not determine *which* chains to use, just how many.
756 */
757static int iwl_get_active_rx_chain_count(struct iwl_priv *priv)
758{
759	if (priv->lib->bt_params &&
760	    priv->lib->bt_params->advanced_bt_coexist &&
761	    (priv->bt_full_concurrent ||
762	     priv->bt_traffic_load >= IWL_BT_COEX_TRAFFIC_LOAD_HIGH)) {
763		/*
764		 * only use chain 'A' in bt high traffic load or
765		 * full concurrency mode
766		 */
767		return IWL_NUM_RX_CHAINS_SINGLE;
768	}
769	/* # of Rx chains to use when expecting MIMO. */
770	if (is_single_rx_stream(priv))
771		return IWL_NUM_RX_CHAINS_SINGLE;
772	else
773		return IWL_NUM_RX_CHAINS_MULTIPLE;
774}
775
776/*
777 * When we are in power saving mode, unless device support spatial
778 * multiplexing power save, use the active count for rx chain count.
779 */
780static int iwl_get_idle_rx_chain_count(struct iwl_priv *priv, int active_cnt)
781{
782	/* # Rx chains when idling, depending on SMPS mode */
783	switch (priv->current_ht_config.smps) {
784	case IEEE80211_SMPS_STATIC:
785	case IEEE80211_SMPS_DYNAMIC:
786		return IWL_NUM_IDLE_CHAINS_SINGLE;
787	case IEEE80211_SMPS_AUTOMATIC:
788	case IEEE80211_SMPS_OFF:
789		return active_cnt;
790	default:
791		WARN(1, "invalid SMPS mode %d",
792		     priv->current_ht_config.smps);
793		return active_cnt;
794	}
795}
796
797/* up to 4 chains */
798static u8 iwl_count_chain_bitmap(u32 chain_bitmap)
799{
800	u8 res;
801	res = (chain_bitmap & BIT(0)) >> 0;
802	res += (chain_bitmap & BIT(1)) >> 1;
803	res += (chain_bitmap & BIT(2)) >> 2;
804	res += (chain_bitmap & BIT(3)) >> 3;
805	return res;
806}
807
808/**
809 * iwlagn_set_rxon_chain - Set up Rx chain usage in "staging" RXON image
810 *
811 * Selects how many and which Rx receivers/antennas/chains to use.
812 * This should not be used for scan command ... it puts data in wrong place.
813 */
814void iwlagn_set_rxon_chain(struct iwl_priv *priv, struct iwl_rxon_context *ctx)
815{
816	bool is_single = is_single_rx_stream(priv);
817	bool is_cam = !test_bit(STATUS_POWER_PMI, &priv->status);
818	u8 idle_rx_cnt, active_rx_cnt, valid_rx_cnt;
819	u32 active_chains;
820	u16 rx_chain;
821
822	/* Tell uCode which antennas are actually connected.
823	 * Before first association, we assume all antennas are connected.
824	 * Just after first association, iwl_chain_noise_calibration()
825	 *    checks which antennas actually *are* connected. */
826	if (priv->chain_noise_data.active_chains)
827		active_chains = priv->chain_noise_data.active_chains;
828	else
829		active_chains = priv->nvm_data->valid_rx_ant;
830
831	if (priv->lib->bt_params &&
832	    priv->lib->bt_params->advanced_bt_coexist &&
833	    (priv->bt_full_concurrent ||
834	     priv->bt_traffic_load >= IWL_BT_COEX_TRAFFIC_LOAD_HIGH)) {
835		/*
836		 * only use chain 'A' in bt high traffic load or
837		 * full concurrency mode
838		 */
839		active_chains = first_antenna(active_chains);
840	}
841
842	rx_chain = active_chains << RXON_RX_CHAIN_VALID_POS;
843
844	/* How many receivers should we use? */
845	active_rx_cnt = iwl_get_active_rx_chain_count(priv);
846	idle_rx_cnt = iwl_get_idle_rx_chain_count(priv, active_rx_cnt);
847
848
849	/* correct rx chain count according hw settings
850	 * and chain noise calibration
851	 */
852	valid_rx_cnt = iwl_count_chain_bitmap(active_chains);
853	if (valid_rx_cnt < active_rx_cnt)
854		active_rx_cnt = valid_rx_cnt;
855
856	if (valid_rx_cnt < idle_rx_cnt)
857		idle_rx_cnt = valid_rx_cnt;
858
859	rx_chain |= active_rx_cnt << RXON_RX_CHAIN_MIMO_CNT_POS;
860	rx_chain |= idle_rx_cnt  << RXON_RX_CHAIN_CNT_POS;
861
862	ctx->staging.rx_chain = cpu_to_le16(rx_chain);
863
864	if (!is_single && (active_rx_cnt >= IWL_NUM_RX_CHAINS_SINGLE) && is_cam)
865		ctx->staging.rx_chain |= RXON_RX_CHAIN_MIMO_FORCE_MSK;
866	else
867		ctx->staging.rx_chain &= ~RXON_RX_CHAIN_MIMO_FORCE_MSK;
868
869	IWL_DEBUG_ASSOC(priv, "rx_chain=0x%X active=%d idle=%d\n",
870			ctx->staging.rx_chain,
871			active_rx_cnt, idle_rx_cnt);
872
873	WARN_ON(active_rx_cnt == 0 || idle_rx_cnt == 0 ||
874		active_rx_cnt < idle_rx_cnt);
875}
876
877u8 iwl_toggle_tx_ant(struct iwl_priv *priv, u8 ant, u8 valid)
878{
879	int i;
880	u8 ind = ant;
881
882	if (priv->band == IEEE80211_BAND_2GHZ &&
883	    priv->bt_traffic_load >= IWL_BT_COEX_TRAFFIC_LOAD_HIGH)
884		return 0;
885
886	for (i = 0; i < RATE_ANT_NUM - 1; i++) {
887		ind = (ind + 1) < RATE_ANT_NUM ?  ind + 1 : 0;
888		if (valid & BIT(ind))
889			return ind;
890	}
891	return ant;
892}
893
894#ifdef CONFIG_PM_SLEEP
895static void iwlagn_convert_p1k(u16 *p1k, __le16 *out)
896{
897	int i;
898
899	for (i = 0; i < IWLAGN_P1K_SIZE; i++)
900		out[i] = cpu_to_le16(p1k[i]);
901}
902
903struct wowlan_key_data {
904	struct iwl_rxon_context *ctx;
905	struct iwlagn_wowlan_rsc_tsc_params_cmd *rsc_tsc;
906	struct iwlagn_wowlan_tkip_params_cmd *tkip;
907	const u8 *bssid;
908	bool error, use_rsc_tsc, use_tkip;
909};
910
911
912static void iwlagn_wowlan_program_keys(struct ieee80211_hw *hw,
913			       struct ieee80211_vif *vif,
914			       struct ieee80211_sta *sta,
915			       struct ieee80211_key_conf *key,
916			       void *_data)
917{
918	struct iwl_priv *priv = IWL_MAC80211_GET_DVM(hw);
919	struct wowlan_key_data *data = _data;
920	struct iwl_rxon_context *ctx = data->ctx;
921	struct aes_sc *aes_sc, *aes_tx_sc = NULL;
922	struct tkip_sc *tkip_sc, *tkip_tx_sc = NULL;
923	struct iwlagn_p1k_cache *rx_p1ks;
924	u8 *rx_mic_key;
925	struct ieee80211_key_seq seq;
926	u32 cur_rx_iv32 = 0;
927	u16 p1k[IWLAGN_P1K_SIZE];
928	int ret, i;
929
930	mutex_lock(&priv->mutex);
931
932	if ((key->cipher == WLAN_CIPHER_SUITE_WEP40 ||
933	     key->cipher == WLAN_CIPHER_SUITE_WEP104) &&
934	     !sta && !ctx->key_mapping_keys)
935		ret = iwl_set_default_wep_key(priv, ctx, key);
936	else
937		ret = iwl_set_dynamic_key(priv, ctx, key, sta);
938
939	if (ret) {
940		IWL_ERR(priv, "Error setting key during suspend!\n");
941		data->error = true;
942	}
943
944	switch (key->cipher) {
945	case WLAN_CIPHER_SUITE_TKIP:
946		if (sta) {
947			tkip_sc = data->rsc_tsc->all_tsc_rsc.tkip.unicast_rsc;
948			tkip_tx_sc = &data->rsc_tsc->all_tsc_rsc.tkip.tsc;
949
950			rx_p1ks = data->tkip->rx_uni;
951
952			ieee80211_get_key_tx_seq(key, &seq);
953			tkip_tx_sc->iv16 = cpu_to_le16(seq.tkip.iv16);
954			tkip_tx_sc->iv32 = cpu_to_le32(seq.tkip.iv32);
955
956			ieee80211_get_tkip_p1k_iv(key, seq.tkip.iv32, p1k);
957			iwlagn_convert_p1k(p1k, data->tkip->tx.p1k);
958
959			memcpy(data->tkip->mic_keys.tx,
960			       &key->key[NL80211_TKIP_DATA_OFFSET_TX_MIC_KEY],
961			       IWLAGN_MIC_KEY_SIZE);
962
963			rx_mic_key = data->tkip->mic_keys.rx_unicast;
964		} else {
965			tkip_sc =
966				data->rsc_tsc->all_tsc_rsc.tkip.multicast_rsc;
967			rx_p1ks = data->tkip->rx_multi;
968			rx_mic_key = data->tkip->mic_keys.rx_mcast;
969		}
970
971		/*
972		 * For non-QoS this relies on the fact that both the uCode and
973		 * mac80211 use TID 0 (as they need to to avoid replay attacks)
974		 * for checking the IV in the frames.
975		 */
976		for (i = 0; i < IWLAGN_NUM_RSC; i++) {
977			ieee80211_get_key_rx_seq(key, i, &seq);
978			tkip_sc[i].iv16 = cpu_to_le16(seq.tkip.iv16);
979			tkip_sc[i].iv32 = cpu_to_le32(seq.tkip.iv32);
980			/* wrapping isn't allowed, AP must rekey */
981			if (seq.tkip.iv32 > cur_rx_iv32)
982				cur_rx_iv32 = seq.tkip.iv32;
983		}
984
985		ieee80211_get_tkip_rx_p1k(key, data->bssid, cur_rx_iv32, p1k);
986		iwlagn_convert_p1k(p1k, rx_p1ks[0].p1k);
987		ieee80211_get_tkip_rx_p1k(key, data->bssid,
988					  cur_rx_iv32 + 1, p1k);
989		iwlagn_convert_p1k(p1k, rx_p1ks[1].p1k);
990
991		memcpy(rx_mic_key,
992		       &key->key[NL80211_TKIP_DATA_OFFSET_RX_MIC_KEY],
993		       IWLAGN_MIC_KEY_SIZE);
994
995		data->use_tkip = true;
996		data->use_rsc_tsc = true;
997		break;
998	case WLAN_CIPHER_SUITE_CCMP:
999		if (sta) {
1000			u8 *pn = seq.ccmp.pn;
1001
1002			aes_sc = data->rsc_tsc->all_tsc_rsc.aes.unicast_rsc;
1003			aes_tx_sc = &data->rsc_tsc->all_tsc_rsc.aes.tsc;
1004
1005			ieee80211_get_key_tx_seq(key, &seq);
1006			aes_tx_sc->pn = cpu_to_le64(
1007					(u64)pn[5] |
1008					((u64)pn[4] << 8) |
1009					((u64)pn[3] << 16) |
1010					((u64)pn[2] << 24) |
1011					((u64)pn[1] << 32) |
1012					((u64)pn[0] << 40));
1013		} else
1014			aes_sc = data->rsc_tsc->all_tsc_rsc.aes.multicast_rsc;
1015
1016		/*
1017		 * For non-QoS this relies on the fact that both the uCode and
1018		 * mac80211 use TID 0 for checking the IV in the frames.
1019		 */
1020		for (i = 0; i < IWLAGN_NUM_RSC; i++) {
1021			u8 *pn = seq.ccmp.pn;
1022
1023			ieee80211_get_key_rx_seq(key, i, &seq);
1024			aes_sc->pn = cpu_to_le64(
1025					(u64)pn[5] |
1026					((u64)pn[4] << 8) |
1027					((u64)pn[3] << 16) |
1028					((u64)pn[2] << 24) |
1029					((u64)pn[1] << 32) |
1030					((u64)pn[0] << 40));
1031		}
1032		data->use_rsc_tsc = true;
1033		break;
1034	}
1035
1036	mutex_unlock(&priv->mutex);
1037}
1038
1039int iwlagn_send_patterns(struct iwl_priv *priv,
1040			struct cfg80211_wowlan *wowlan)
1041{
1042	struct iwlagn_wowlan_patterns_cmd *pattern_cmd;
1043	struct iwl_host_cmd cmd = {
1044		.id = REPLY_WOWLAN_PATTERNS,
1045		.dataflags[0] = IWL_HCMD_DFL_NOCOPY,
1046	};
1047	int i, err;
1048
1049	if (!wowlan->n_patterns)
1050		return 0;
1051
1052	cmd.len[0] = sizeof(*pattern_cmd) +
1053		wowlan->n_patterns * sizeof(struct iwlagn_wowlan_pattern);
1054
1055	pattern_cmd = kmalloc(cmd.len[0], GFP_KERNEL);
1056	if (!pattern_cmd)
1057		return -ENOMEM;
1058
1059	pattern_cmd->n_patterns = cpu_to_le32(wowlan->n_patterns);
1060
1061	for (i = 0; i < wowlan->n_patterns; i++) {
1062		int mask_len = DIV_ROUND_UP(wowlan->patterns[i].pattern_len, 8);
1063
1064		memcpy(&pattern_cmd->patterns[i].mask,
1065			wowlan->patterns[i].mask, mask_len);
1066		memcpy(&pattern_cmd->patterns[i].pattern,
1067			wowlan->patterns[i].pattern,
1068			wowlan->patterns[i].pattern_len);
1069		pattern_cmd->patterns[i].mask_size = mask_len;
1070		pattern_cmd->patterns[i].pattern_size =
1071			wowlan->patterns[i].pattern_len;
1072	}
1073
1074	cmd.data[0] = pattern_cmd;
1075	err = iwl_dvm_send_cmd(priv, &cmd);
1076	kfree(pattern_cmd);
1077	return err;
1078}
1079
1080int iwlagn_suspend(struct iwl_priv *priv, struct cfg80211_wowlan *wowlan)
1081{
1082	struct iwlagn_wowlan_wakeup_filter_cmd wakeup_filter_cmd;
1083	struct iwl_rxon_cmd rxon;
1084	struct iwl_rxon_context *ctx = &priv->contexts[IWL_RXON_CTX_BSS];
1085	struct iwlagn_wowlan_kek_kck_material_cmd kek_kck_cmd;
1086	struct iwlagn_wowlan_tkip_params_cmd tkip_cmd = {};
1087	struct iwlagn_d3_config_cmd d3_cfg_cmd = {
1088		/*
1089		 * Program the minimum sleep time to 10 seconds, as many
1090		 * platforms have issues processing a wakeup signal while
1091		 * still being in the process of suspending.
1092		 */
1093		.min_sleep_time = cpu_to_le32(10 * 1000 * 1000),
1094	};
1095	struct wowlan_key_data key_data = {
1096		.ctx = ctx,
1097		.bssid = ctx->active.bssid_addr,
1098		.use_rsc_tsc = false,
1099		.tkip = &tkip_cmd,
1100		.use_tkip = false,
1101	};
1102	int ret, i;
1103	u16 seq;
1104
1105	key_data.rsc_tsc = kzalloc(sizeof(*key_data.rsc_tsc), GFP_KERNEL);
1106	if (!key_data.rsc_tsc)
1107		return -ENOMEM;
1108
1109	memset(&wakeup_filter_cmd, 0, sizeof(wakeup_filter_cmd));
1110
1111	/*
1112	 * We know the last used seqno, and the uCode expects to know that
1113	 * one, it will increment before TX.
1114	 */
1115	seq = le16_to_cpu(priv->last_seq_ctl) & IEEE80211_SCTL_SEQ;
1116	wakeup_filter_cmd.non_qos_seq = cpu_to_le16(seq);
1117
1118	/*
1119	 * For QoS counters, we store the one to use next, so subtract 0x10
1120	 * since the uCode will add 0x10 before using the value.
1121	 */
1122	for (i = 0; i < IWL_MAX_TID_COUNT; i++) {
1123		seq = priv->tid_data[IWL_AP_ID][i].seq_number;
1124		seq -= 0x10;
1125		wakeup_filter_cmd.qos_seq[i] = cpu_to_le16(seq);
1126	}
1127
1128	if (wowlan->disconnect)
1129		wakeup_filter_cmd.enabled |=
1130			cpu_to_le32(IWLAGN_WOWLAN_WAKEUP_BEACON_MISS |
1131				    IWLAGN_WOWLAN_WAKEUP_LINK_CHANGE);
1132	if (wowlan->magic_pkt)
1133		wakeup_filter_cmd.enabled |=
1134			cpu_to_le32(IWLAGN_WOWLAN_WAKEUP_MAGIC_PACKET);
1135	if (wowlan->gtk_rekey_failure)
1136		wakeup_filter_cmd.enabled |=
1137			cpu_to_le32(IWLAGN_WOWLAN_WAKEUP_GTK_REKEY_FAIL);
1138	if (wowlan->eap_identity_req)
1139		wakeup_filter_cmd.enabled |=
1140			cpu_to_le32(IWLAGN_WOWLAN_WAKEUP_EAP_IDENT_REQ);
1141	if (wowlan->four_way_handshake)
1142		wakeup_filter_cmd.enabled |=
1143			cpu_to_le32(IWLAGN_WOWLAN_WAKEUP_4WAY_HANDSHAKE);
1144	if (wowlan->n_patterns)
1145		wakeup_filter_cmd.enabled |=
1146			cpu_to_le32(IWLAGN_WOWLAN_WAKEUP_PATTERN_MATCH);
1147
1148	if (wowlan->rfkill_release)
1149		d3_cfg_cmd.wakeup_flags |=
1150			cpu_to_le32(IWLAGN_D3_WAKEUP_RFKILL);
1151
1152	iwl_scan_cancel_timeout(priv, 200);
1153
1154	memcpy(&rxon, &ctx->active, sizeof(rxon));
1155
1156	priv->ucode_loaded = false;
1157	iwl_trans_stop_device(priv->trans);
1158
1159	priv->wowlan = true;
1160
1161	ret = iwl_load_ucode_wait_alive(priv, IWL_UCODE_WOWLAN);
1162	if (ret)
1163		goto out;
1164
1165	/* now configure WoWLAN ucode */
1166	ret = iwl_alive_start(priv);
1167	if (ret)
1168		goto out;
1169
1170	memcpy(&ctx->staging, &rxon, sizeof(rxon));
1171	ret = iwlagn_commit_rxon(priv, ctx);
1172	if (ret)
1173		goto out;
1174
1175	ret = iwl_power_update_mode(priv, true);
1176	if (ret)
1177		goto out;
1178
1179	if (!iwlwifi_mod_params.sw_crypto) {
1180		/* mark all keys clear */
1181		priv->ucode_key_table = 0;
1182		ctx->key_mapping_keys = 0;
1183
1184		/*
1185		 * This needs to be unlocked due to lock ordering
1186		 * constraints. Since we're in the suspend path
1187		 * that isn't really a problem though.
1188		 */
1189		mutex_unlock(&priv->mutex);
1190		ieee80211_iter_keys(priv->hw, ctx->vif,
1191				    iwlagn_wowlan_program_keys,
1192				    &key_data);
1193		mutex_lock(&priv->mutex);
1194		if (key_data.error) {
1195			ret = -EIO;
1196			goto out;
1197		}
1198
1199		if (key_data.use_rsc_tsc) {
1200			struct iwl_host_cmd rsc_tsc_cmd = {
1201				.id = REPLY_WOWLAN_TSC_RSC_PARAMS,
1202				.data[0] = key_data.rsc_tsc,
1203				.dataflags[0] = IWL_HCMD_DFL_NOCOPY,
1204				.len[0] = sizeof(*key_data.rsc_tsc),
1205			};
1206
1207			ret = iwl_dvm_send_cmd(priv, &rsc_tsc_cmd);
1208			if (ret)
1209				goto out;
1210		}
1211
1212		if (key_data.use_tkip) {
1213			ret = iwl_dvm_send_cmd_pdu(priv,
1214						 REPLY_WOWLAN_TKIP_PARAMS,
1215						 0, sizeof(tkip_cmd),
1216						 &tkip_cmd);
1217			if (ret)
1218				goto out;
1219		}
1220
1221		if (priv->have_rekey_data) {
1222			memset(&kek_kck_cmd, 0, sizeof(kek_kck_cmd));
1223			memcpy(kek_kck_cmd.kck, priv->kck, NL80211_KCK_LEN);
1224			kek_kck_cmd.kck_len = cpu_to_le16(NL80211_KCK_LEN);
1225			memcpy(kek_kck_cmd.kek, priv->kek, NL80211_KEK_LEN);
1226			kek_kck_cmd.kek_len = cpu_to_le16(NL80211_KEK_LEN);
1227			kek_kck_cmd.replay_ctr = priv->replay_ctr;
1228
1229			ret = iwl_dvm_send_cmd_pdu(priv,
1230						 REPLY_WOWLAN_KEK_KCK_MATERIAL,
1231						 0, sizeof(kek_kck_cmd),
1232						 &kek_kck_cmd);
1233			if (ret)
1234				goto out;
1235		}
1236	}
1237
1238	ret = iwl_dvm_send_cmd_pdu(priv, REPLY_D3_CONFIG, 0,
1239				     sizeof(d3_cfg_cmd), &d3_cfg_cmd);
1240	if (ret)
1241		goto out;
1242
1243	ret = iwl_dvm_send_cmd_pdu(priv, REPLY_WOWLAN_WAKEUP_FILTER,
1244				 0, sizeof(wakeup_filter_cmd),
1245				 &wakeup_filter_cmd);
1246	if (ret)
1247		goto out;
1248
1249	ret = iwlagn_send_patterns(priv, wowlan);
1250 out:
1251	kfree(key_data.rsc_tsc);
1252	return ret;
1253}
1254#endif
1255
1256int iwl_dvm_send_cmd(struct iwl_priv *priv, struct iwl_host_cmd *cmd)
1257{
1258	if (iwl_is_rfkill(priv) || iwl_is_ctkill(priv)) {
1259		IWL_WARN(priv, "Not sending command - %s KILL\n",
1260			 iwl_is_rfkill(priv) ? "RF" : "CT");
1261		return -EIO;
1262	}
1263
1264	if (test_bit(STATUS_FW_ERROR, &priv->status)) {
1265		IWL_ERR(priv, "Command %s failed: FW Error\n",
1266			iwl_dvm_get_cmd_string(cmd->id));
1267		return -EIO;
1268	}
1269
1270	/*
1271	 * This can happen upon FW ASSERT: we clear the STATUS_FW_ERROR flag
1272	 * in iwl_down but cancel the workers only later.
1273	 */
1274	if (!priv->ucode_loaded) {
1275		IWL_ERR(priv, "Fw not loaded - dropping CMD: %x\n", cmd->id);
1276		return -EIO;
1277	}
1278
1279	/*
1280	 * Synchronous commands from this op-mode must hold
1281	 * the mutex, this ensures we don't try to send two
1282	 * (or more) synchronous commands at a time.
1283	 */
1284	if (!(cmd->flags & CMD_ASYNC))
1285		lockdep_assert_held(&priv->mutex);
1286
1287	return iwl_trans_send_cmd(priv->trans, cmd);
1288}
1289
1290int iwl_dvm_send_cmd_pdu(struct iwl_priv *priv, u8 id,
1291			 u32 flags, u16 len, const void *data)
1292{
1293	struct iwl_host_cmd cmd = {
1294		.id = id,
1295		.len = { len, },
1296		.data = { data, },
1297		.flags = flags,
1298	};
1299
1300	return iwl_dvm_send_cmd(priv, &cmd);
1301}
1302